mirror of
https://github.com/amd/blis.git
synced 2026-04-24 09:38:51 +00:00
Implemented and registered power9 dgemm ukernel. Details: - Implemented 12x6 dgemm microkernel for power9. This microkernel assumes that elements of B have been duplicated/broadcast during the packing step. The microkernel uses a column orientation for its microtile vector registers and thus implements column storage and general stride IO cases. (A row storage IO case via in-register transposition may be added at a future date.) It should be noted that we recommend using this microkernel with gcc and *not* xlc, as issues with the latter cropped up during development, including but not limited to slightly incompatible vector register mnemonics in the GNU extended inline assembly clobber list.
149 lines
6.3 KiB
C
149 lines
6.3 KiB
C
/*
|
|
|
|
BLIS
|
|
An object-based framework for developing high-performance BLAS-like
|
|
libraries.
|
|
|
|
Copyright (C) 2014, The University of Texas at Austin
|
|
Copyright (C) 2019, Advanced Micro Devices, Inc.
|
|
|
|
Redistribution and use in source and binary forms, with or without
|
|
modification, are permitted provided that the following conditions are
|
|
met:
|
|
- Redistributions of source code must retain the above copyright
|
|
notice, this list of conditions and the following disclaimer.
|
|
- Redistributions in binary form must reproduce the above copyright
|
|
notice, this list of conditions and the following disclaimer in the
|
|
documentation and/or other materials provided with the distribution.
|
|
- Neither the name(s) of the copyright holder(s) nor the names of its
|
|
contributors may be used to endorse or promote products derived
|
|
from this software without specific prior written permission.
|
|
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
// -- level-3 ------------------------------------------------------------------
|
|
|
|
// gemm (asm d6x8)
|
|
GEMM_UKR_PROT( float, s, gemm_haswell_asm_6x16 )
|
|
GEMM_UKR_PROT( double, d, gemm_haswell_asm_6x8 )
|
|
GEMM_UKR_PROT( scomplex, c, gemm_haswell_asm_3x8 )
|
|
GEMM_UKR_PROT( dcomplex, z, gemm_haswell_asm_3x4 )
|
|
|
|
// gemm (asm d8x6)
|
|
GEMM_UKR_PROT( float, s, gemm_haswell_asm_16x6 )
|
|
GEMM_UKR_PROT( double, d, gemm_haswell_asm_8x6 )
|
|
GEMM_UKR_PROT( scomplex, c, gemm_haswell_asm_8x3 )
|
|
GEMM_UKR_PROT( dcomplex, z, gemm_haswell_asm_4x3 )
|
|
|
|
// gemmtrsm_l (asm d6x8)
|
|
GEMMTRSM_UKR_PROT( float, s, gemmtrsm_l_haswell_asm_6x16 )
|
|
GEMMTRSM_UKR_PROT( double, d, gemmtrsm_l_haswell_asm_6x8 )
|
|
|
|
// gemmtrsm_u (asm d6x8)
|
|
GEMMTRSM_UKR_PROT( float, s, gemmtrsm_u_haswell_asm_6x16 )
|
|
GEMMTRSM_UKR_PROT( double, d, gemmtrsm_u_haswell_asm_6x8 )
|
|
|
|
|
|
// gemm (asm d8x6)
|
|
//GEMM_UKR_PROT( float, s, gemm_haswell_asm_16x6 )
|
|
//GEMM_UKR_PROT( double, d, gemm_haswell_asm_8x6 )
|
|
//GEMM_UKR_PROT( scomplex, c, gemm_haswell_asm_8x3 )
|
|
//GEMM_UKR_PROT( dcomplex, z, gemm_haswell_asm_4x3 )
|
|
|
|
|
|
// -- level-3 sup --------------------------------------------------------------
|
|
|
|
// gemmsup_rv
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_6x8 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_5x8 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_4x8 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_3x8 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_2x8 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_1x8 )
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_6x6 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_5x6 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_4x6 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_3x6 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_2x6 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_1x6 )
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_6x4 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_5x4 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_4x4 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_3x4 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_2x4 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_1x4 )
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_6x2 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_5x2 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_4x2 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_3x2 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_2x2 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_1x2 )
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_r_haswell_ref_6x1 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_r_haswell_ref_5x1 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_r_haswell_ref_4x1 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_r_haswell_ref_3x1 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_r_haswell_ref_2x1 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_r_haswell_ref_1x1 )
|
|
|
|
// gemmsup_rv (mkernel in m dim)
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_6x8m )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_6x6m )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_6x4m )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_6x2m )
|
|
|
|
// gemmsup_rv (mkernel in n dim)
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_6x8n )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_5x8n )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_4x8n )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_3x8n )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_2x8n )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rv_haswell_asm_1x8n )
|
|
|
|
// gemmsup_rd
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_6x8 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_2x8 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_1x8 )
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_6x4 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_2x4 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_1x4 )
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_6x2 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_3x2 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_2x2 )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_1x2 )
|
|
|
|
// gemmsup_rd (mkernel in m dim)
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_6x8m )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_6x4m )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_6x2m )
|
|
|
|
// gemmsup_rd (mkernel in n dim)
|
|
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_6x8n )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_3x8n )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_2x8n )
|
|
GEMMSUP_KER_PROT( double, d, gemmsup_rd_haswell_asm_1x8n )
|
|
|