Anthony Chang
9287b7c6b3
Grouped batched attention + permute (#412)
* grouped attn without batch validates; now move toward grouped batched attn
* grouped batched attention
* working
* remove debug logging
clean up
clean up
* reintroduce g_ prefix back to host tensor variables
* format
* rename file
* restore old file
* rename
* consolidate padded/non-padded attention example
* harmonize padding specialization in attn examples
2022-09-19 16:09:44 -05:00
..
2022-09-01 09:31:17 -05:00
2022-07-29 18:19:25 -05:00
2022-07-29 18:19:25 -05:00
2022-08-23 10:38:41 -05:00
2022-09-01 09:31:17 -05:00
2022-08-31 16:32:17 -05:00
2022-08-25 16:58:48 -05:00
2022-07-29 18:19:25 -05:00
2022-07-29 18:19:25 -05:00
2022-08-25 17:19:15 -05:00
2022-08-30 11:38:26 -05:00
2022-07-29 18:19:25 -05:00
2022-08-13 00:16:14 -05:00
2022-08-15 10:11:02 -05:00
2022-07-29 18:19:25 -05:00
2022-08-15 10:11:02 -05:00
2022-08-25 17:19:15 -05:00
2022-09-06 12:22:48 -05:00
2022-08-25 17:19:15 -05:00
2022-08-12 15:22:39 -05:00
2022-07-29 18:19:25 -05:00
2022-08-13 09:43:18 -05:00
2022-08-10 12:20:29 -05:00
2022-08-10 12:20:29 -05:00
2022-09-19 11:25:28 -05:00
2022-09-01 09:31:17 -05:00
2022-09-19 16:09:44 -05:00
2022-08-15 10:11:02 -05:00
2022-08-15 10:11:02 -05:00
2022-08-25 17:19:15 -05:00
2022-09-09 10:41:15 -05:00
2022-09-14 17:54:18 -05:00
2022-09-19 11:25:28 -05:00
2022-09-01 09:31:17 -05:00
2022-09-19 11:25:28 -05:00