mirror of
https://github.com/amd/blis.git
synced 2026-05-04 06:21:12 +00:00
* Updated aocl_batch_gemm_ APIs aligning to CBLAS batch API. - Modified Batch-Gemm API to align with cblas_?gemm_batch_ API, and added a parameter group_size to the existing APIs. - Updated bench batch_gemm code to align to the new API definition. - Modified the hardcoded number in lpgemm_postop file. - Added necessary early return condition to account for group_count/group_size < 0. AMD-Internal: [ SWLCSG - 3592 ]
236 lines
8.5 KiB
C
236 lines
8.5 KiB
C
/*
|
|
|
|
BLIS
|
|
An object-based framework for developing high-performance BLAS-like
|
|
libraries.
|
|
|
|
Copyright (C) 2022 - 2025, Advanced Micro Devices, Inc. All rights reserved.
|
|
|
|
Redistribution and use in source and binary forms, with or without
|
|
modification, are permitted provided that the following conditions are
|
|
met:
|
|
- Redistributions of source code must retain the above copyright
|
|
notice, this list of conditions and the following disclaimer.
|
|
- Redistributions in binary form must reproduce the above copyright
|
|
notice, this list of conditions and the following disclaimer in the
|
|
documentation and/or other materials provided with the distribution.
|
|
- Neither the name(s) of the copyright holder(s) nor the names of its
|
|
contributors may be used to endorse or promote products derived
|
|
from this software without specific prior written permission.
|
|
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
#ifndef AOCL_GEMM_INTERFACE_H
|
|
#define AOCL_GEMM_INTERFACE_H
|
|
|
|
#include "lpgemm_types.h"
|
|
#include "aocl_bf16_type.h"
|
|
|
|
// Returns the size of buffer in bytes required for the reordered matrix.
|
|
#define AOCL_GEMM_GET_REORDER_BUF_SIZE(LP_SFX) \
|
|
BLIS_EXPORT_ADDON siz_t aocl_get_reorder_buf_size_ ## LP_SFX \
|
|
( \
|
|
const char order, \
|
|
const char trans, \
|
|
const char mat_type, \
|
|
const dim_t k, \
|
|
const dim_t n \
|
|
) \
|
|
|
|
AOCL_GEMM_GET_REORDER_BUF_SIZE(f32f32f32of32);
|
|
AOCL_GEMM_GET_REORDER_BUF_SIZE(u8s8s32os32);
|
|
AOCL_GEMM_GET_REORDER_BUF_SIZE(bf16bf16f32of32);
|
|
AOCL_GEMM_GET_REORDER_BUF_SIZE(s8s8s32os32);
|
|
AOCL_GEMM_GET_REORDER_BUF_SIZE(u8s4s32os32);
|
|
AOCL_GEMM_GET_REORDER_BUF_SIZE(bf16s4f32of32);
|
|
|
|
// Returns the size of buffer in bytes required for the reordered matrix.
|
|
#define AOCL_GEMM_GET_REORDER_BUF_SIZE_SYM_QUANT(LP_SFX) \
|
|
BLIS_EXPORT_ADDON siz_t aocl_get_reorder_buf_size_ ## LP_SFX \
|
|
( \
|
|
const char order, \
|
|
const char trans, \
|
|
const char mat_type, \
|
|
const dim_t k, \
|
|
const dim_t n, \
|
|
AOCL_SYMM_STAT_QUANT* meta_data \
|
|
) \
|
|
|
|
AOCL_GEMM_GET_REORDER_BUF_SIZE_SYM_QUANT(s8s8s32os32_sym_quant);
|
|
|
|
// Performs reordering of input matrix. Reordering is the process of packing
|
|
// the entire matrix upfront, so that the benefits of packed matrix is obtained
|
|
// without incurring the packing costs during matmul computation.
|
|
#define AOCL_GEMM_REORDER(B_type,LP_SFX) \
|
|
BLIS_EXPORT_ADDON void aocl_reorder_ ## LP_SFX \
|
|
( \
|
|
const char order, \
|
|
const char trans, \
|
|
const char mat_type, \
|
|
const B_type* input_buf_addr, \
|
|
B_type* reorder_buf_addr, \
|
|
const dim_t k, \
|
|
const dim_t n, \
|
|
const dim_t ldb \
|
|
) \
|
|
|
|
AOCL_GEMM_REORDER(float,f32f32f32of32);
|
|
AOCL_GEMM_REORDER(int8_t,u8s8s32os32);
|
|
AOCL_GEMM_REORDER(bfloat16,bf16bf16f32of32);
|
|
AOCL_GEMM_REORDER(bfloat16,bf16bf16f32of32_reference);
|
|
AOCL_GEMM_REORDER(int8_t,s8s8s32os32);
|
|
AOCL_GEMM_REORDER(int8_t,u8s4s32os32);
|
|
AOCL_GEMM_REORDER(int8_t, bf16s4f32of32);
|
|
|
|
#define AOCL_GEMM_REORDER_SYM_QUANT(B_type,LP_SFX) \
|
|
BLIS_EXPORT_ADDON void aocl_reorder_ ## LP_SFX \
|
|
( \
|
|
const char order, \
|
|
const char trans, \
|
|
const char mat_type, \
|
|
const B_type* input_buf_addr, \
|
|
B_type* reorder_buf_addr, \
|
|
const dim_t k, \
|
|
const dim_t n, \
|
|
const dim_t ldb, \
|
|
AOCL_SYMM_STAT_QUANT* meta_data \
|
|
) \
|
|
|
|
AOCL_GEMM_REORDER_SYM_QUANT(int8_t,s8s8s32os32_sym_quant);
|
|
|
|
#define AOCL_GEMM_REORDER_MXP(A_type,B_type,LP_SFX) \
|
|
BLIS_EXPORT_ADDON void aocl_reorder_ ## LP_SFX \
|
|
( \
|
|
const char order, \
|
|
const char trans, \
|
|
const char mat_type, \
|
|
const A_type* input_buf_addr, \
|
|
B_type* reorder_buf_addr, \
|
|
const dim_t k, \
|
|
const dim_t n, \
|
|
const dim_t ldb \
|
|
) \
|
|
|
|
AOCL_GEMM_REORDER_MXP(float,bfloat16,f32obf16);
|
|
|
|
#define AOCL_GEMM_UNREORDER(B_type, LP_SFX) \
|
|
BLIS_EXPORT_ADDON void aocl_unreorder_ ## LP_SFX \
|
|
( \
|
|
const char order, \
|
|
const char mat_type, \
|
|
const B_type* reorder_buf_addr, \
|
|
B_type* output_buf_addr, \
|
|
const dim_t k, \
|
|
const dim_t n, \
|
|
const dim_t ldb \
|
|
) \
|
|
|
|
AOCL_GEMM_UNREORDER(bfloat16, bf16bf16f32of32);
|
|
AOCL_GEMM_UNREORDER(bfloat16, bf16bf16f32of32_reference);
|
|
AOCL_GEMM_UNREORDER(int8_t, s8s8s32os32_reference);
|
|
|
|
#define AOCL_GEMM_MATMUL(A_type,B_type,C_type,Sum_type,LP_SFX) \
|
|
BLIS_EXPORT_ADDON void aocl_gemm_ ## LP_SFX \
|
|
( \
|
|
const char order, \
|
|
const char transa, \
|
|
const char transb, \
|
|
const dim_t m, \
|
|
const dim_t n, \
|
|
const dim_t k, \
|
|
const Sum_type alpha, \
|
|
const A_type* a, \
|
|
const dim_t lda, \
|
|
const char mem_format_a, \
|
|
const B_type* b, \
|
|
const dim_t ldb, \
|
|
const char mem_format_b, \
|
|
const Sum_type beta, \
|
|
C_type* c, \
|
|
const dim_t ldc, \
|
|
aocl_post_op* post_op_unparsed \
|
|
) \
|
|
|
|
AOCL_GEMM_MATMUL(uint8_t,int8_t,int32_t,int32_t,u8s8s32os32);
|
|
AOCL_GEMM_MATMUL(uint8_t,int8_t,int8_t,int32_t,u8s8s32os8);
|
|
AOCL_GEMM_MATMUL(uint8_t,int8_t,bfloat16,int32_t,u8s8s32obf16);
|
|
AOCL_GEMM_MATMUL(uint8_t,int8_t,float,int32_t,u8s8s32of32);
|
|
AOCL_GEMM_MATMUL(uint8_t,int8_t,uint8_t,int32_t,u8s8s32ou8);
|
|
|
|
AOCL_GEMM_MATMUL(int8_t,int8_t,int32_t,int32_t,s8s8s32os32);
|
|
AOCL_GEMM_MATMUL(int8_t,int8_t,int8_t,int32_t,s8s8s32os8);
|
|
AOCL_GEMM_MATMUL(int8_t,int8_t,bfloat16,int32_t,s8s8s32obf16);
|
|
AOCL_GEMM_MATMUL(int8_t,int8_t,float,int32_t,s8s8s32of32);
|
|
AOCL_GEMM_MATMUL(int8_t,int8_t,uint8_t,int32_t,s8s8s32ou8);
|
|
|
|
// Symmetric static quantization GEMM API
|
|
AOCL_GEMM_MATMUL(int8_t,int8_t,float,int32_t,s8s8s32of32_sym_quant);
|
|
AOCL_GEMM_MATMUL(int8_t,int8_t,bfloat16,int32_t,s8s8s32obf16_sym_quant);
|
|
|
|
AOCL_GEMM_MATMUL(bfloat16,bfloat16,bfloat16,float,bf16bf16f32obf16);
|
|
AOCL_GEMM_MATMUL(bfloat16,bfloat16,float,float,bf16bf16f32of32);
|
|
AOCL_GEMM_MATMUL(bfloat16, int8_t, float, float, bf16s4f32of32);
|
|
AOCL_GEMM_MATMUL(bfloat16, int8_t, bfloat16, float, bf16s4f32obf16);
|
|
|
|
AOCL_GEMM_MATMUL(float,float,float,float,f32f32f32of32);
|
|
|
|
#define AOCL_BGEMM_MATMUL(A_type,B_type,C_type,Sum_type,LP_SFX) \
|
|
BLIS_EXPORT_ADDON void aocl_batch_gemm_ ## LP_SFX \
|
|
( \
|
|
const char* order, \
|
|
const char* transa, \
|
|
const char* transb, \
|
|
const dim_t* m, \
|
|
const dim_t* n, \
|
|
const dim_t* k, \
|
|
const Sum_type* alpha, \
|
|
const A_type** a, \
|
|
const dim_t* lda, \
|
|
const B_type** b, \
|
|
const dim_t* ldb, \
|
|
const Sum_type* beta, \
|
|
C_type** c, \
|
|
const dim_t* ldc, \
|
|
const dim_t group_count, \
|
|
const dim_t* group_size, \
|
|
const char* mem_format_a, \
|
|
const char* mem_format_b, \
|
|
aocl_post_op** post_op_unparsed \
|
|
) \
|
|
|
|
// bf16 APIs
|
|
AOCL_BGEMM_MATMUL(bfloat16,bfloat16,float,float,bf16bf16f32of32);
|
|
AOCL_BGEMM_MATMUL(bfloat16,bfloat16,bfloat16,float,bf16bf16f32obf16);
|
|
AOCL_BGEMM_MATMUL(bfloat16,int8_t,float,float,bf16s4f32of32);
|
|
AOCL_BGEMM_MATMUL(bfloat16,int8_t,bfloat16,float,bf16s4f32obf16);
|
|
// f32 APIs
|
|
AOCL_BGEMM_MATMUL(float,float,float,float,f32f32f32of32);
|
|
// u8s8 APIs
|
|
AOCL_BGEMM_MATMUL(uint8_t,int8_t,int32_t,int32_t,u8s8s32os32);
|
|
AOCL_BGEMM_MATMUL(uint8_t,int8_t,int8_t,int32_t,u8s8s32os8);
|
|
AOCL_BGEMM_MATMUL(uint8_t,int8_t,float,int32_t,u8s8s32of32);
|
|
AOCL_BGEMM_MATMUL(uint8_t,int8_t,bfloat16,int32_t,u8s8s32obf16);
|
|
AOCL_BGEMM_MATMUL(uint8_t,int8_t,uint8_t,int32_t,u8s8s32ou8);
|
|
// s8s8 APIs
|
|
AOCL_BGEMM_MATMUL(int8_t,int8_t,int32_t,int32_t,s8s8s32os32);
|
|
AOCL_BGEMM_MATMUL(int8_t,int8_t,int8_t,int32_t,s8s8s32os8);
|
|
AOCL_BGEMM_MATMUL(int8_t,int8_t,float,int32_t,s8s8s32of32);
|
|
AOCL_BGEMM_MATMUL(int8_t,int8_t,bfloat16,int32_t,s8s8s32obf16);
|
|
AOCL_BGEMM_MATMUL(int8_t,int8_t,uint8_t,int32_t,s8s8s32ou8);
|
|
|
|
|
|
|
|
#endif // AOCL_GEMM_INTERFACE_H
|