Anthony Chang
9287b7c6b3
Grouped batched attention + permute (#412)
* grouped attn without batch validates; now move toward grouped batched attn
* grouped batched attention
* working
* remove debug logging
clean up
clean up
* reintroduce g_ prefix back to host tensor variables
* format
* rename file
* restore old file
* rename
* consolidate padded/non-padded attention example
* harmonize padding specialization in attn examples
2022-09-19 16:09:44 -05:00
..
2022-09-19 16:09:44 -05:00
2022-08-15 10:11:02 -05:00
2022-08-15 10:11:02 -05:00
2022-06-24 23:32:43 -05:00
2022-06-24 23:32:43 -05:00
2022-09-08 09:27:50 -05:00
2022-09-14 17:54:18 -05:00
2022-09-08 09:27:50 -05:00
2022-06-24 23:32:43 -05:00
2022-08-15 10:11:02 -05:00
2022-06-27 14:25:10 -05:00
2022-06-24 23:32:43 -05:00
2022-06-24 23:32:43 -05:00
2022-06-24 23:32:43 -05:00
2022-06-24 23:32:43 -05:00
2022-08-24 10:12:54 -05:00
2022-09-19 11:25:28 -05:00
2022-06-24 23:32:43 -05:00
2022-07-08 15:55:14 -05:00
2022-06-27 14:25:10 -05:00
2022-07-08 15:55:14 -05:00
2022-07-01 01:38:00 -05:00
2022-06-24 23:32:43 -05:00
2022-08-13 09:18:58 -05:00
2022-06-24 23:32:43 -05:00
2022-06-24 23:32:43 -05:00
2022-08-23 14:41:56 -05:00
2022-06-24 23:32:43 -05:00
2022-06-24 23:32:43 -05:00
2022-06-24 23:32:43 -05:00
2022-08-13 09:43:18 -05:00
2022-08-13 09:43:18 -05:00
2022-06-24 23:32:43 -05:00
2022-08-15 10:11:02 -05:00
2022-07-14 22:52:45 -05:00
2022-09-09 10:41:15 -05:00