Files
blis/addon/aocl_gemm/config/lpgemm_config.c
Mithun Mohan 097cda9f9e Adding support for AOCL_ENABLE_INSTRUCTIONS for f32 LPGEMM API.
-Currently lpgemm sets the context (block sizes and micro-kernels) based
on the ISA of the machine it is being executed on. However this approach
does not give the flexibility to select a different context at runtime.
In order to enable runtime selection of context, the context
initialization is modified to read the AOCL_ENABLE_INSTRUCTIONS env
variable and set the context based on the same. As part of this commit,
only f32 context selection is enabled.
-Bug fixes in scale ops in f32 micro-kernels and GEMV path selection.
-Added vectorized f32 packing kernels for NR=16(AVX2) and NR=64(AVX512).
This is only for B matrix and helps remove dependency of f32 lpgemm api
on the BLIS packing framework.

AMD Internal: [CPUPL-5959]

Change-Id: I4b459aaf33c54423952f89905ba43cf119ce20f6
2024-10-30 08:52:22 +00:00

519 lines
14 KiB
C

/*
BLIS
An object-based framework for developing high-performance BLAS-like
libraries.
Copyright (C) 2022 - 2024, Advanced Micro Devices, Inc. All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
- Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
- Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in the
documentation and/or other materials provided with the distribution.
- Neither the name(s) of the copyright holder(s) nor the names of its
contributors may be used to endorse or promote products derived
from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "blis.h"
#include "lpgemm_config.h"
#include "lpgemm_func_map.h"
#include "lpgemm_blksz_map.h"
#include "lpgemm_kernels.h"
#include "lpgemm_pack_bf16.h"
#include "lpgemm_packb_s16.h"
#include "lpgemm_packa_s16.h"
#include "lpgemm_packa.h"
#include "lpgemm_packb.h"
#include "lpgemm_packa_s8.h"
#include "lpgemm_packb_s8.h"
#include "lpgemm_packb_s8s16.h"
#include "lpgemm_pack_f32.h"
static lpgemm_cntx_t global_cntx_t_list[AOCL_OPERATION_TYPE_LEN] \
__attribute__((aligned(64))); //Only one op type supported now.
static lpgemm_util_cntx_t global_util_cntx_t_list[AOCL_UTIL_OPERATION_TYPE_LEN] \
__attribute__((aligned(64))); //Only post-ops like utils.
static lpgemm_eltwise_ops_cntx_t
global_eltwise_ops_cntx_t_list[AOCL_ELTWISE_OPS_OPERATION_TYPE_LEN] \
__attribute__((aligned(64))); //Post-ops only utils without gemm.
static arch_t global_lpgemm_enable_arch = BLIS_ARCH_ERROR;
// This array is to store function pointers to jit generated kernels.
static void* global_jit_kernels[ LPGEMM_BF16_MR ]
[ ( LPGEMM_BF16_NR / NUM_F32_ELEMS_PER_ZMM ) + 1 ]
__attribute__((aligned(64)));
// Buffer size is chosen in order to accommodate the
// worst-case scenario for MR=6 and NR=64.
// The buffersize is chosen using bruteforce method.
#define JIT_KERNEL_SIZE ( 10 * BLIS_PAGE_SIZE )
static bli_pthread_once_t once_check_lpgemm_func_map_init = BLIS_PTHREAD_ONCE_INIT;
static void _lpgemm_init_enable_arch()
{
arch_t arch_id = bli_arch_query_id();
bool enbl_instr = bli_aocl_enable_instruction_query();
if ( ( enbl_instr == TRUE ) &&
( ( arch_id == BLIS_ARCH_ZEN3 ) ||
( arch_id == BLIS_ARCH_ZEN2 ) ||
( arch_id == BLIS_ARCH_ZEN ) ) )
{
global_lpgemm_enable_arch = BLIS_ARCH_ZEN3;
}
}
arch_t lpgemm_get_enabled_arch()
{
return global_lpgemm_enable_arch;
}
static void _lpgemm_util_cntx_init_func_map()
{
#define UMACRO(ID,FUNC_PTR) global_util_cntx_t_list[ID].kern_fun_ptr = FUNC_PTR;
global_util_cntx_t_list[F32_GELU_TANH].kern_fun_ptr = NULL;
global_util_cntx_t_list[F32_GELU_ERF].kern_fun_ptr = NULL;
global_util_cntx_t_list[F32_SOFTMAX].kern_fun_ptr = NULL;
// Kernel dispatch object factory.
if ( bli_cpuid_is_avx512bf16_supported() == TRUE )
{
#ifdef BLIS_KERNELS_ZEN4
LPGEMM_UTIL_KERN_FUNC_MAP_AVX512_VNNI_BF16
#endif
}
else if ( bli_cpuid_is_avx512vnni_supported() == TRUE )
{
#ifdef BLIS_KERNELS_ZEN4
LPGEMM_UTIL_KERN_FUNC_MAP_AVX512_VNNI
#endif
}
else if ( bli_cpuid_is_avx2fma3_supported() == TRUE )
{
#ifdef BLIS_KERNELS_ZEN3
LPGEMM_UTIL_KERN_FUNC_MAP_AVX2
#endif
}
#undef UMACRO
}
static void _lpgemm_eltwise_ops_cntx_init_func_map()
{
#define POMACRO(ID,FUNC_PTR) \
global_eltwise_ops_cntx_t_list[ID].eltwise_ops_kern_fun_ptr = FUNC_PTR;
global_eltwise_ops_cntx_t_list[BF16OF32].eltwise_ops_kern_fun_ptr = NULL;
// Kernel dispatch object factory.
if ( bli_cpuid_is_avx512bf16_supported() == TRUE )
{
#ifdef BLIS_KERNELS_ZEN4
LPGEMM_ELTWISE_OPS_KERN_FUNC_MAP_AVX512_VNNI_BF16
#endif
}
#undef POMACRO
}
static void _lpgemm_cntx_init_func_map()
{
#define KMACRO(ID,FUNC_PTR) global_cntx_t_list[ID].kern_fun_ptr = FUNC_PTR;
#define PAMACRO(ID,FUNC_PTR) global_cntx_t_list[ID].packa_fun_ptr = FUNC_PTR;
#define PBMACRO(ID,FUNC_PTR) global_cntx_t_list[ID].packb_fun_ptr = FUNC_PTR;
#define UBMACRO(ID, FUNC_PTR) global_cntx_t_list[ID].unpackb_fun_ptr = FUNC_PTR;
#define PBSMACRO(ID, FUNC_PTR) global_cntx_t_list[ID].packsclb_fun_ptr = FUNC_PTR;
#define JITMACRO(ID, FUNC_PTR) global_cntx_t_list[ID].jit_kernel = FUNC_PTR;
//TODO: Default initialize with reference kernels so that kernel pointer
// will be valid even in case none of the zen optimized kernels are
// available. This scenario could happen if the addon was built using
// a different arch config (eg: skx).
global_cntx_t_list[U8S8S16OS16].kern_fun_ptr = NULL;
global_cntx_t_list[U8S8S32OS32].kern_fun_ptr = NULL;
global_cntx_t_list[F32F32F32OF32].kern_fun_ptr = NULL;
global_cntx_t_list[BF16BF16F32OF32].kern_fun_ptr = NULL;
global_cntx_t_list[BF16S4F32OF32].kern_fun_ptr = NULL;
// Kernel dispatch object factory.
if ( bli_cpuid_is_avx512bf16_supported() == TRUE )
{
#ifdef BLIS_KERNELS_ZEN4
LPGEMM_KERN_FUNC_MAP_AVX512_VNNI_BF16
LPGEMM_PACKA_FUNC_MAP_AVX512_VNNI_BF16
LPGEMM_PACKB_FUNC_MAP_AVX512_VNNI_BF16
LPGEMM_UNPACKB_FUNC_MAP_AVX512_VNNI_BF16
LPGEMM_PACKSCLB_FUNC_MAP_AVX512_VNNI_BF16
#ifdef LPGEMM_BF16_JIT
lpgemm_jit_inputs_t inputs;
inputs.alpha_scale = TRUE;
inputs.beta_scale = BLIS_BETA_GEN;
err_t err;
dim_t num_N_vars = ( LPGEMM_BF16_NR / NUM_F32_ELEMS_PER_ZMM ) + 1;
for ( dim_t m = 0; m < LPGEMM_BF16_MR; m++ )
{
for( dim_t n = 0; n < num_N_vars; n++ )
{
inputs.MR = ( m == 0 ) ? LPGEMM_BF16_MR : m;
inputs.NR = n * 16;
inputs.m_loop = ( m == 0 ) ? TRUE: FALSE;
inputs.generate_mask = ( n == 0 ) ? TRUE: FALSE;
global_jit_kernels[m][n] = bli_malloc_user( JIT_KERNEL_SIZE,
&err );
if( global_jit_kernels[m][n] != NULL )
{
get_jit_kernel( &inputs,
global_jit_kernels[m][n],
JIT_KERNEL_SIZE
);
}
}
}
#endif
#endif
// If arch is updated at runtime, it is expeceted to be honoured.
if ( global_lpgemm_enable_arch == BLIS_ARCH_ZEN3 )
{
LPGEMM_KERN_FUNC_UPD_MAP_AVX512_VNNI_BF16_TO_AVX2;
LPGEMM_PACKB_FUNC_UPD_MAP_AVX512_VNNI_BF16_TO_AVX2;
}
}
else if ( bli_cpuid_is_avx512vnni_supported() == TRUE )
{
#ifdef BLIS_KERNELS_ZEN4
LPGEMM_KERN_FUNC_MAP_AVX512_VNNI
LPGEMM_PACKA_FUNC_MAP_AVX512_VNNI
LPGEMM_PACKB_FUNC_MAP_AVX512_VNNI
#endif
if ( global_lpgemm_enable_arch == BLIS_ARCH_ZEN3 )
{
LPGEMM_KERN_FUNC_UPD_MAP_AVX512_VNNI_TO_AVX2
LPGEMM_PACKB_FUNC_UPD_MAP_AVX512_VNNI_TO_AVX2;
}
}
else if ( bli_cpuid_is_avx2fma3_supported() == TRUE )
{
#ifdef BLIS_KERNELS_ZEN3
LPGEMM_KERN_FUNC_MAP_AVX2
LPGEMM_PACKA_FUNC_MAP_AVX2
LPGEMM_PACKB_FUNC_MAP_AVX2
#endif
}
// If built with a config not supporting zen3/zen4/amdzen, error out
// since reference kernels are not available.
if ( global_cntx_t_list[F32F32F32OF32].kern_fun_ptr == NULL )
{
bli_print_msg( "AOCL_GEMM is not compiled using correct Zen config."
" Compile using zen3/zen4/amdzen config.",
__FILE__, __LINE__ );
bli_abort();
}
#undef PBMACRO
#undef PAMACRO
#undef KMACRO
}
void lpgemm_set_jit_kernel( void* kernel_fp, dim_t m_index, dim_t n_index )
{
global_jit_kernels[m_index][n_index] = kernel_fp;
}
void* lpgemm_get_jit_kernel( dim_t m_index, dim_t n_index )
{
return global_jit_kernels[m_index][n_index];
}
BLIS_INLINE void lpgemm_set_block_sizes_global_cntx
(
AOCL_OPERATION_TYPE op_type,
dim_t MC,
dim_t NC,
dim_t KC,
dim_t MR,
dim_t NR
)
{
global_cntx_t_list[op_type].blksz.MC = MC;
global_cntx_t_list[op_type].blksz.NC = NC;
global_cntx_t_list[op_type].blksz.KC = KC;
global_cntx_t_list[op_type].blksz.MR = MR;
global_cntx_t_list[op_type].blksz.NR = NR;
}
BLIS_INLINE void lpgemm_set_pack_strides_global_cntx
(
AOCL_OPERATION_TYPE op_type,
dim_t packa_rs,
dim_t packa_cs,
dim_t packb_rs,
dim_t packb_cs
)
{
global_cntx_t_list[op_type].pack_s.packa_rs = packa_rs;
global_cntx_t_list[op_type].pack_s.packa_cs = packa_cs;
global_cntx_t_list[op_type].pack_s.packb_rs = packb_rs;
global_cntx_t_list[op_type].pack_s.packb_cs = packb_cs;
}
static void _lpgemm_cntx_init_blksz_map()
{
#define XMACRO(ID,MC,NC,KC,MR,NR,PACKA_RS,PACKA_CS,PACKB_RS,PACKB_CS) \
lpgemm_set_block_sizes_global_cntx(ID, MC, NC, KC, MR, NR); \
lpgemm_set_pack_strides_global_cntx(ID, PACKA_RS, PACKA_CS, PACKB_RS, PACKB_CS);
// Ideally the blocksize needs to be set based on arch id. However
// since this code is also expected to work on other vendor machines,
// the blocksize for a particular version of zen id is generalized
// for all machines that support the ISA supported by that particular
// zen id.
if ( bli_cpuid_is_avx512vnni_supported() == TRUE )
{
LPGEMM_BLKSZ_MAP_ZEN4
if ( global_lpgemm_enable_arch == BLIS_ARCH_ZEN3 )
{
LPGEMM_BLKSZ_UPD_MAP_ZEN4_TO_ZEN
}
}
else if ( bli_cpuid_is_avx2fma3_supported() == TRUE )
{
LPGEMM_BLKSZ_MAP_ZEN
}
else
{
LPGEMM_BLKSZ_MAP_ZEN
}
#undef XMACRO
}
BLIS_INLINE void lpgemm_set_sup_thres_global_cntx
(
AOCL_OPERATION_TYPE op_type,
dim_t MT,
dim_t NT,
dim_t KT
)
{
global_cntx_t_list[op_type].sup_thres.MT = MT;
global_cntx_t_list[op_type].sup_thres.NT = NT;
global_cntx_t_list[op_type].sup_thres.KT = KT;
}
static void _lpgemm_cntx_init_sup_thres_map()
{
#define STMACRO(ID,MT,NT,KT) \
lpgemm_set_sup_thres_global_cntx(ID, MT, NT, KT); \
if ( bli_cpuid_is_avx512vnni_supported() == TRUE )
{
LPGEMM_SUP_THRES_MAP_ZEN4
if ( global_lpgemm_enable_arch == BLIS_ARCH_ZEN3 )
{
LPGEMM_SUP_THRES_UPD_MAP_ZEN4_TO_ZEN
}
}
else if ( bli_cpuid_is_avx2fma3_supported() == TRUE )
{
LPGEMM_SUP_THRES_MAP_ZEN
}
else
{
LPGEMM_SUP_THRES_MAP_ZEN
}
#undef STMACRO
}
BLIS_INLINE void lpgemm_set_block_sizes_global_eltwise_ops_cntx
(
AOCL_ELTWISE_OPS_OPERATION_TYPE op_type,
dim_t MC,
dim_t NC,
dim_t KC,
dim_t MR,
dim_t NR
)
{
global_eltwise_ops_cntx_t_list[op_type].blksz.MC = MC;
global_eltwise_ops_cntx_t_list[op_type].blksz.NC = NC;
global_eltwise_ops_cntx_t_list[op_type].blksz.KC = KC;
global_eltwise_ops_cntx_t_list[op_type].blksz.MR = MR;
global_eltwise_ops_cntx_t_list[op_type].blksz.NR = NR;
}
static void _lpgemm_eltwise_ops_cntx_init_blksz_map()
{
#define XMACRO(ID,MC,NC,KC,MR,NR) \
lpgemm_set_block_sizes_global_eltwise_ops_cntx(ID, MC, NC, KC, MR, NR);
// Ideally the blocksize needs to be set based on arch id. However
// since this code is also expected to work on other vendor machines,
// the blocksize for a particular version of zen id is generalized
// for all machines that support the ISA supported by that particular
// zen id.
if ( bli_cpuid_is_avx512bf16_supported() == TRUE )
{
LPGEMM_ELTWISE_OPS_BLKSZ_MAP_ZEN4
}
else
{
LPGEMM_ELTWISE_OPS_BLKSZ_MAP_ZEN
}
#undef XMACRO
}
static void lpgemm_cntx_init_map()
{
_lpgemm_init_enable_arch();
_lpgemm_cntx_init_func_map();
_lpgemm_cntx_init_blksz_map();
_lpgemm_cntx_init_sup_thres_map();
_lpgemm_eltwise_ops_cntx_init_blksz_map();
_lpgemm_eltwise_ops_cntx_init_func_map();
_lpgemm_util_cntx_init_func_map();
}
// Sets default block sizes for lpgemm. Currently only u8s8s32 supported.
void aocl_lpgemm_init_global_cntx()
{
bli_pthread_once
(
&once_check_lpgemm_func_map_init,
lpgemm_cntx_init_map
);
}
lpgemm_cntx_t* lpgemm_get_global_cntx_obj( AOCL_OPERATION_TYPE op )
{
return &global_cntx_t_list[op];
}
lpgemm_util_cntx_t* lpgemm_util_get_global_cntx_obj( AOCL_UTIL_OPERATION_TYPE op )
{
return &global_util_cntx_t_list[op];
}
lpgemm_eltwise_ops_cntx_t* lpgemm_eltwise_ops_get_global_cntx_obj
( AOCL_ELTWISE_OPS_OPERATION_TYPE op )
{
return &global_eltwise_ops_cntx_t_list[op];
}
dim_t lpgemm_get_block_size_MC_global_cntx( AOCL_OPERATION_TYPE op_type )
{
return global_cntx_t_list[op_type].blksz.MC;
}
dim_t lpgemm_get_block_size_NC_global_cntx( AOCL_OPERATION_TYPE op_type )
{
return global_cntx_t_list[op_type].blksz.NC;
}
dim_t lpgemm_get_block_size_KC_global_cntx( AOCL_OPERATION_TYPE op_type )
{
return global_cntx_t_list[op_type].blksz.KC;
}
dim_t lpgemm_get_block_size_NR_global_cntx( AOCL_OPERATION_TYPE op_type )
{
return global_cntx_t_list[op_type].blksz.NR;
}
dim_t lpgemm_get_block_size_MR_global_cntx( AOCL_OPERATION_TYPE op_type )
{
return global_cntx_t_list[op_type].blksz.MR;
}
dim_t lpgemm_get_sup_thres_MT_global_cntx( AOCL_OPERATION_TYPE op_type )
{
return global_cntx_t_list[op_type].sup_thres.MT;
}
dim_t lpgemm_get_sup_thres_NT_global_cntx( AOCL_OPERATION_TYPE op_type )
{
return global_cntx_t_list[op_type].sup_thres.NT;
}
dim_t lpgemm_get_sup_thres_KT_global_cntx( AOCL_OPERATION_TYPE op_type )
{
return global_cntx_t_list[op_type].sup_thres.KT;
}
void lpgemm_get_packa_strides( lpgemm_cntx_t* lcntx, dim_t* rs, dim_t* cs )
{
*rs = lcntx->pack_s.packa_rs;
*cs = lcntx->pack_s.packa_cs;
}
void lpgemm_get_packb_strides( lpgemm_cntx_t* lcntx, dim_t* rs, dim_t* cs )
{
*rs = lcntx->pack_s.packb_rs;
*cs = lcntx->pack_s.packb_cs;
}
void lpgemm_mod_block_size_s16
(
dim_t m,
dim_t n,
dim_t k,
dim_t* MC,
dim_t* NC,
dim_t* KC
)
{
const dim_t range[4] = {1024, 512, 256, 128};
if (n < *NC)
{
for (dim_t i = 0; i < 4; ++i)
{
if (n <= range[i])
{
*NC = range[i];
}
}
}
if (k < *KC)
{
for (dim_t i = 0; i < 4; ++i)
{
if (k <= range[i])
{
*KC = range[i];
}
}
}
}