From a95f64601d6d49c72da5be638b30d8009fac685c Mon Sep 17 00:00:00 2001 From: Qianfeng Zhang Date: Tue, 7 Apr 2026 09:38:18 +0000 Subject: [PATCH] Remove dropout=true instances to reduce compiling-time --- .../18_hstu_attention/generate_instances.py | 42 +-- .../hstu_attention_group_forward_bf16.cpp | 9 +- .../hstu_attention_group_forward_fp16.cpp | 9 +- .../hstu_attention_no_group_forward_bf16.cpp | 11 +- .../hstu_attention_no_group_forward_fp16.cpp | 11 +- ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...ion_batched_forward_bf16_instances_ref.hpp | 256 ------------------ ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...ion_batched_forward_fp16_instances_ref.hpp | 256 ------------------ ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...ntion_group_forward_bf16_instances_ref.hpp | 256 ------------------ ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...ntion_group_forward_fp16_instances_ref.hpp | 256 ------------------ ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...tion_jagged_forward_bf16_instances_ref.hpp | 256 ------------------ ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- ...tion_jagged_forward_fp16_instances_ref.hpp | 256 ------------------ ...ax_false_has_bias_has_dropout_maxk_128.cpp | 18 -- ...ax_false_has_bias_has_dropout_maxk_256.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_64.cpp | 18 -- ...max_false_has_bias_has_dropout_maxk_96.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_128.cpp | 18 -- ...max_false_no_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_false_no_bias_has_dropout_maxk_96.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_128.cpp | 18 -- ...max_true_has_bias_has_dropout_maxk_256.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_64.cpp | 18 -- ...tmax_true_has_bias_has_dropout_maxk_96.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_128.cpp | 18 -- ...tmax_true_no_bias_has_dropout_maxk_256.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_64.cpp | 18 -- ...ftmax_true_no_bias_has_dropout_maxk_96.cpp | 18 -- 203 files changed, 42 insertions(+), 5032 deletions(-) delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp delete mode 100644 example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp diff --git a/example/ck_tile/18_hstu_attention/generate_instances.py b/example/ck_tile/18_hstu_attention/generate_instances.py index e27432e879..11375c1293 100644 --- a/example/ck_tile/18_hstu_attention/generate_instances.py +++ b/example/ck_tile/18_hstu_attention/generate_instances.py @@ -82,19 +82,22 @@ MODE_GROUP_OR_NOT_MAP = { "group": "Group", } + def create_forward_instances(instance_dir: Path, headdims: List) -> None: for mode in ["batched", "jagged", "group"]: for dtype in ["fp16", "bf16"]: for has_causal in [True, False]: for use_softmax in [True, False]: for has_bias in [True, False]: - for has_dropout in [True, False]: + for has_dropout in [False]: for max_k in headdims: fname = HSTU_FORWARD_INSTANCE_FNAME.format( mode=mode, dtype_str=dtype, has_or_no_causal_str=BOOL_MAP_CAUSAL[has_causal], - use_softmax_or_not_str=BOOL_MAP_SOFTMAX[use_softmax], + use_softmax_or_not_str=BOOL_MAP_SOFTMAX[ + use_softmax + ], has_or_no_bias_str=BOOL_MAP_BIAS[has_bias], has_or_no_dropout_str=BOOL_MAP_DROPOUT[has_dropout], max_k_str=INT_MAP_MAX_K[max_k], @@ -105,22 +108,24 @@ def create_forward_instances(instance_dir: Path, headdims: List) -> None: dtype_file=TYPE_FNAME_MAP[dtype], ) ) - forward_instance = HSTU_FORWARD_INSTANCE_TEMPLATE.format( - extern="", - mode=mode, - dtype=TYPE_CTYPE_MAP[dtype], - has_causal=BOOL_MAP[has_causal], - use_softmax=BOOL_MAP[use_softmax], - has_bias=BOOL_MAP[has_bias], - has_dropout=BOOL_MAP[has_dropout], - max_k=max_k, - group_or_not=MODE_GROUP_OR_NOT_MAP[mode], + forward_instance = ( + HSTU_FORWARD_INSTANCE_TEMPLATE.format( + extern="", + mode=mode, + dtype=TYPE_CTYPE_MAP[dtype], + has_causal=BOOL_MAP[has_causal], + use_softmax=BOOL_MAP[use_softmax], + has_bias=BOOL_MAP[has_bias], + has_dropout=BOOL_MAP[has_dropout], + max_k=max_k, + group_or_not=MODE_GROUP_OR_NOT_MAP[mode], + ) ) (instance_dir / fname).write_text( - HSTU_COPYRIGHT_HEADER - + forward_instance_inc - + forward_instance - ) + HSTU_COPYRIGHT_HEADER + + forward_instance_inc + + forward_instance + ) def create_forward_instances_ref(instance_dir: Path, headdims: List) -> None: @@ -141,7 +146,7 @@ def create_forward_instances_ref(instance_dir: Path, headdims: List) -> None: file.write(forward_instance_inc) for max_k in headdims: for has_bias in [True, False]: - for has_dropout in [True, False]: + for has_dropout in [False]: for has_causal in [True, False]: for use_softmax in [True, False]: forward_instance = ( @@ -156,9 +161,10 @@ def create_forward_instances_ref(instance_dir: Path, headdims: List) -> None: max_k=max_k, group_or_not=MODE_GROUP_OR_NOT_MAP[mode], ) - ) + ) file.write(forward_instance) + if __name__ == "__main__": headdims_fwd = [64, 96, 128, 256] diff --git a/example/ck_tile/18_hstu_attention/hstu_attention_group_forward_bf16.cpp b/example/ck_tile/18_hstu_attention/hstu_attention_group_forward_bf16.cpp index c9e55d8b9f..9748de693d 100644 --- a/example/ck_tile/18_hstu_attention/hstu_attention_group_forward_bf16.cpp +++ b/example/ck_tile/18_hstu_attention/hstu_attention_group_forward_bf16.cpp @@ -12,17 +12,16 @@ void hstu_attention_group_forward_bf16(HstuAttentionGroupFwdParams& param, hipStream_t stream) { - const bool has_dropout = (param.p_drop > 0.0f); - const bool has_bias = (param.bias_ptr != nullptr); - const bool use_causal = param.use_causal; - BOOL_SWITCH_3(has_bias, kHasBias, has_dropout, kHasDropout, use_causal, kUseCausal, [&] { + const bool has_bias = (param.bias_ptr != nullptr); + const bool use_causal = param.use_causal; + BOOL_SWITCH_2(has_bias, kHasBias, use_causal, kUseCausal, [&] { HDIM_SWITCH(param.hdim_qk, param.hdim_v, MaxK, [&] { BOOL_SWITCH(param.use_softmax, kUseSoftmax, [&] { run_group_forward_causal_softmax_bias_dropout_dispatch(param, stream); }); }); diff --git a/example/ck_tile/18_hstu_attention/hstu_attention_group_forward_fp16.cpp b/example/ck_tile/18_hstu_attention/hstu_attention_group_forward_fp16.cpp index 858ef4c268..14d473499a 100644 --- a/example/ck_tile/18_hstu_attention/hstu_attention_group_forward_fp16.cpp +++ b/example/ck_tile/18_hstu_attention/hstu_attention_group_forward_fp16.cpp @@ -12,18 +12,17 @@ void hstu_attention_group_forward_fp16(HstuAttentionGroupFwdParams& param, hipStream_t stream) { - const bool has_dropout = (param.p_drop > 0.0f); - const bool has_bias = (param.bias_ptr != nullptr); - const bool use_causal = param.use_causal; + const bool has_bias = (param.bias_ptr != nullptr); + const bool use_causal = param.use_causal; - BOOL_SWITCH_3(has_bias, kHasBias, has_dropout, kHasDropout, use_causal, kUseCausal, [&] { + BOOL_SWITCH_2(has_bias, kHasBias, use_causal, kUseCausal, [&] { HDIM_SWITCH(param.hdim_qk, param.hdim_v, MaxK, [&] { BOOL_SWITCH(param.use_softmax, kUseSoftmax, [&] { run_group_forward_causal_softmax_bias_dropout_dispatch(param, stream); }); }); diff --git a/example/ck_tile/18_hstu_attention/hstu_attention_no_group_forward_bf16.cpp b/example/ck_tile/18_hstu_attention/hstu_attention_no_group_forward_bf16.cpp index d0f9164a3b..b2e4dfd20b 100644 --- a/example/ck_tile/18_hstu_attention/hstu_attention_no_group_forward_bf16.cpp +++ b/example/ck_tile/18_hstu_attention/hstu_attention_no_group_forward_bf16.cpp @@ -14,10 +14,9 @@ void hstu_attention_no_group_forward_bf16(HstuAttentionNoGroupFwdParams& param, hipStream_t stream) { - const bool has_dropout = (param.p_drop > 0.0f); - const bool has_bias = (param.bias_ptr != nullptr); - const bool use_causal = param.use_causal; - BOOL_SWITCH_3(has_bias, kHasBias, has_dropout, kHasDropout, use_causal, kUseCausal, [&] { + const bool has_bias = (param.bias_ptr != nullptr); + const bool use_causal = param.use_causal; + BOOL_SWITCH_2(has_bias, kHasBias, use_causal, kUseCausal, [&] { HDIM_SWITCH(param.hdim_qk, param.hdim_v, MaxK, [&] { BOOL_SWITCH(param.use_softmax, kUseSoftmax, [&] { if(param.is_jagged) @@ -25,14 +24,14 @@ void hstu_attention_no_group_forward_bf16(HstuAttentionNoGroupFwdParams& param, kUseCausal, kUseSoftmax, kHasBias, - kHasDropout, + false, // kHasDropout MaxK>(param, stream); else run_batched_forward_causal_softmax_bias_dropout_dispatch(param, stream); }); }); diff --git a/example/ck_tile/18_hstu_attention/hstu_attention_no_group_forward_fp16.cpp b/example/ck_tile/18_hstu_attention/hstu_attention_no_group_forward_fp16.cpp index 5d75873e95..1d4793cf7b 100644 --- a/example/ck_tile/18_hstu_attention/hstu_attention_no_group_forward_fp16.cpp +++ b/example/ck_tile/18_hstu_attention/hstu_attention_no_group_forward_fp16.cpp @@ -14,10 +14,9 @@ void hstu_attention_no_group_forward_fp16(HstuAttentionNoGroupFwdParams& param, hipStream_t stream) { - const bool has_dropout = (param.p_drop > 0.0f); - const bool has_bias = (param.bias_ptr != nullptr); - const bool use_causal = param.use_causal; - BOOL_SWITCH_3(has_bias, kHasBias, has_dropout, kHasDropout, use_causal, kUseCausal, [&] { + const bool has_bias = (param.bias_ptr != nullptr); + const bool use_causal = param.use_causal; + BOOL_SWITCH_2(has_bias, kHasBias, use_causal, kUseCausal, [&] { HDIM_SWITCH(param.hdim_qk, param.hdim_v, MaxK, [&] { BOOL_SWITCH(param.use_softmax, kUseSoftmax, [&] { if(param.is_jagged) @@ -25,14 +24,14 @@ void hstu_attention_no_group_forward_fp16(HstuAttentionNoGroupFwdParams& param, kUseCausal, kUseSoftmax, kHasBias, - kHasDropout, + false, // kHasDropout MaxK>(param, stream); else run_batched_forward_causal_softmax_bias_dropout_dispatch(param, stream); }); }); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 422dda5b4a..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index d1903ae6e7..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 7a05a2be22..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index ac00d7468f..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 2441e1907e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 42fa146645..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 9428dcd4a0..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 9fbb8c5e18..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 0f9dcd42af..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index a69c39fad0..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 0a3b51fcc8..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index ac4aefe9c3..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 5fe4f880d2..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index e34c634e05..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 0460c5b80b..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 2e869aa9f7..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_instances_ref.hpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_instances_ref.hpp index f790245241..44f22df8d2 100644 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_instances_ref.hpp +++ b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_instances_ref.hpp @@ -9,38 +9,6 @@ #include #include "hstu_attention_batched_forward_dispatch.hpp" -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -73,38 +41,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -137,38 +73,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -201,38 +105,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -265,38 +137,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -329,38 +169,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -393,38 +201,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -457,38 +233,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 735e739c83..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index bb61c28263..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index c31ffe1242..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 0a725cff45..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index a4431b7ccb..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 963ca17140..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 7af16744a3..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 796585922e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index a511c9734c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 56f72f67aa..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index ac66d7f6ea..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 8d1e52a03c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 5f394a3ddb..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index df038cf33a..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 8cb42923be..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 8ea33f3364..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index d9d303b485..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 189d7bc817..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 289cee1245..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 3cab083876..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index e683fbd36c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index f046c9197f..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index a50599a974..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 7c9f5bd187..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 0016431f73..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 6493425133..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index f2d2458664..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 38455a0738..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 3c7f2b845c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 18334fd1db..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index c70c74c7b4..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 5b6ab76f4a..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_instances_ref.hpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_instances_ref.hpp index 19111a0713..7037b7aad1 100644 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_instances_ref.hpp +++ b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_instances_ref.hpp @@ -9,38 +9,6 @@ #include #include "hstu_attention_batched_forward_dispatch.hpp" -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -73,38 +41,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -137,38 +73,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -201,38 +105,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -265,38 +137,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -329,38 +169,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -393,38 +201,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -457,38 +233,6 @@ extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< false, 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_batched_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 2121cf92f6..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index dc6b011aef..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 373bb27e59..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 12b4e71783..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 98ab869237..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 494f3a774d..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index dc15fa2451..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index e1cb58bc5c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index f55e6d785d..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 43117f2c55..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 8a36f0f0d9..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index dffb1086bd..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index d2b12a6c21..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 73d922c858..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index e739095d09..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 1182c110bf..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_batched_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_batched_forward_dispatch.hpp" - -template void run_batched_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 72bfdcab90..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 6257e17b69..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index e8d3298767..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 0947fde8ab..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 3339c87dcf..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 9119b5782e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 29fcb28f78..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index c076d7a775..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index f538c1571a..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 8aec6a29a6..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index b4a87b816c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 1bec282dd1..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 23157f2a66..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 3f6d7a9bce..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index f0d34dd7a0..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 215aeafb40..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_instances_ref.hpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_instances_ref.hpp index 04c2d9ebc1..c77f14242a 100644 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_instances_ref.hpp +++ b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_instances_ref.hpp @@ -9,38 +9,6 @@ #include #include "hstu_attention_group_forward_dispatch.hpp" -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -73,38 +41,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -137,38 +73,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -201,38 +105,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -265,38 +137,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -329,38 +169,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -393,38 +201,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -457,38 +233,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index c6ea8ae0b5..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index c073c076f8..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 31000dc322..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 4bd0a083db..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 9fee7a4ab4..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index f644989c06..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 1152616b60..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 14361276ea..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index c063ed8a29..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index f4f41d2482..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 7871e54495..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 61c0c1fca7..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 5cc2377791..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 3bf82440a7..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 10000e7545..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index e783dad3a9..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 34d79f2f09..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index db59a69073..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index ca3d36730a..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 3138e12d78..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index b2626571c0..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 0d1a4d6de8..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index de2eac519f..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index c070c7eeac..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 4b8d3be00c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 7c58121e0f..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 3ce4e17c5e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index f93ed023da..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 8b3b9e206e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index e526c132c3..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 48615d260d..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 4578bad8bb..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_instances_ref.hpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_instances_ref.hpp index baf6e357f0..5bb513f7af 100644 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_instances_ref.hpp +++ b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_instances_ref.hpp @@ -9,38 +9,6 @@ #include #include "hstu_attention_group_forward_dispatch.hpp" -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -73,38 +41,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -137,38 +73,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -201,38 +105,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -265,38 +137,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -329,38 +169,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -393,38 +201,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -457,38 +233,6 @@ extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< false, 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - -extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); - extern template void run_group_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 33774731ce..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index ffe40dc6b0..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 1d7299e0a7..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 7f00f18216..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 5a7e5aa18c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 2969a2512b..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index fe4cb16e0b..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 72752ea8ba..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index af37d59bdc..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 8ce887f9ef..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 4a6d98ce4c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 58caba4598..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index a4459eac56..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 128>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 1860db60b3..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 256>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index eba187323f..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 64>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 9d0928c5ab..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_group_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_group_forward_dispatch.hpp" - -template void run_group_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 96>(HstuAttentionGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 11679e9abc..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 4558f71aec..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 9b1482bc14..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index aeda9deb96..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 3c23ff63ef..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 5646cfbf71..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 25d802246d..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 37f5587895..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 4ada591f5d..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 20b0b15289..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 8b6b85d621..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index fef2cb6120..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index fd64b80b6f..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 8a205d4c72..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 90d560f3a0..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 0e2fb8a008..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_instances_ref.hpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_instances_ref.hpp index cec6bba188..c8329edfa6 100644 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_instances_ref.hpp +++ b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_instances_ref.hpp @@ -9,38 +9,6 @@ #include #include "hstu_attention_jagged_forward_dispatch.hpp" -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -73,38 +41,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -137,38 +73,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -201,38 +105,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -265,38 +137,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -329,38 +169,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -393,38 +201,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, @@ -457,38 +233,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - true, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::bf16_t, true, diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 017b2b9ac2..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 0e0917c32d..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 6e5c7f5c1e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 1257af294e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index a21b46bb60..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index d19e1532be..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 30177b2410..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 70865faa2e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 0cb475ad9e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index ba492409a7..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 664fff3f79..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index bc292bf93e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 003ffabc99..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index ff33c263d2..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 68ea7eea8c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 04780471a3..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_bf16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::bf16_t, - false, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index da56f0e9bc..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index b05e192905..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index baf3f6c201..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index e719a5c212..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 1856b0fa2b..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index bb38021898..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index e6d1a8a6d4..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 74bf15b0b1..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 8d10da5a8e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 0daf202a0d..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index ff0b6a1591..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index cf5872611e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index b71cd4f958..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 6de1ef6209..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 3b89713e20..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 9e9afa6ca7..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_has_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_instances_ref.hpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_instances_ref.hpp index 7d47966fc2..9d7e4bc951 100644 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_instances_ref.hpp +++ b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_instances_ref.hpp @@ -9,38 +9,6 @@ #include #include "hstu_attention_jagged_forward_dispatch.hpp" -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -73,38 +41,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -137,38 +73,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -201,38 +105,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -265,38 +137,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -329,38 +169,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -393,38 +201,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, @@ -457,38 +233,6 @@ extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< false, 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - true, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - -extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); - extern template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< ck_tile::fp16_t, true, diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index f85078a607..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 459499541e..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index dbf21f949a..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index dbec38b765..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index f2e1afc218..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 0b59f1e562..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 4e8fd6401c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 85f52dd708..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_false_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - false, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 5b99069797..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index dc988443f2..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 7c94cf09f9..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 9ecedf3ec8..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_has_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - true, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp deleted file mode 100644 index 7d73b7953c..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_128.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 128>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp deleted file mode 100644 index 430208865b..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_256.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 256>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp deleted file mode 100644 index 67458e2fc5..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_64.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 64>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream); diff --git a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp b/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp deleted file mode 100644 index 8bab4dc559..0000000000 --- a/example/ck_tile/18_hstu_attention/instances/hstu_attention_jagged_forward_fp16_no_causal_softmax_true_no_bias_has_dropout_maxk_96.cpp +++ /dev/null @@ -1,18 +0,0 @@ - -// SPDX-License-Identifier: MIT -// Copyright (c) 2018-2025, Advanced Micro Devices, Inc. All rights reserved. - -// The file is automatically generated, don't modify! -// See the generator script -// `composable_kernel/example/ck_tile/18_hstu_attention/generate_instances.py` - -#include -#include "hstu_attention_jagged_forward_dispatch.hpp" - -template void run_jagged_forward_causal_softmax_bias_dropout_dispatch< - ck_tile::fp16_t, - false, - true, - false, - true, - 96>(HstuAttentionNoGroupFwdParams& param, hipStream_t stream);