mirror of
https://github.com/ikawrakow/ik_llama.cpp.git
synced 2026-04-30 03:11:51 +00:00
q8_k_r16: q2_K and q3_K now uses q8_k_r16 on Zen4+
This commit is contained in:
@@ -2064,14 +2064,21 @@ typedef struct {
|
|||||||
} block_q8_1_r8;
|
} block_q8_1_r8;
|
||||||
|
|
||||||
void iqk_convert_q2_k_q8_k_r8(int n, const void * vx, size_t bx, void * vy, int nrc_x) {
|
void iqk_convert_q2_k_q8_k_r8(int n, const void * vx, size_t bx, void * vy, int nrc_x) {
|
||||||
|
#ifdef HAVE_FANCY_SIMD
|
||||||
|
constexpr int k_nr = 16;
|
||||||
|
using block_q8_k_r = block_q8_k_r16;
|
||||||
|
#else
|
||||||
|
constexpr int k_nr = 8;
|
||||||
|
using block_q8_k_r = block_q8_k_r8;
|
||||||
|
#endif
|
||||||
GGML_ASSERT(n%QK_K == 0);
|
GGML_ASSERT(n%QK_K == 0);
|
||||||
GGML_ASSERT(nrc_x%8 == 0);
|
GGML_ASSERT(nrc_x%k_nr == 0);
|
||||||
|
|
||||||
int nb = n/QK_K;
|
int nb = n/QK_K;
|
||||||
|
|
||||||
const block_q2_K * x8[8];
|
const block_q2_K * x8[k_nr];
|
||||||
|
|
||||||
block_q8_k_r8 * y = (block_q8_k_r8 *)vy;
|
block_q8_k_r * y = (block_q8_k_r *)vy;
|
||||||
|
|
||||||
float f_values[QK_K];
|
float f_values[QK_K];
|
||||||
uint32_t block[8];
|
uint32_t block[8];
|
||||||
@@ -2082,10 +2089,10 @@ void iqk_convert_q2_k_q8_k_r8(int n, const void * vx, size_t bx, void * vy, int
|
|||||||
auto sign_bit = _mm256_set1_ps(-0.0f);
|
auto sign_bit = _mm256_set1_ps(-0.0f);
|
||||||
auto perm = _mm256_setr_epi32(0, 4, 1, 5, 2, 6, 3, 7);
|
auto perm = _mm256_setr_epi32(0, 4, 1, 5, 2, 6, 3, 7);
|
||||||
|
|
||||||
for (int ix = 0; ix < nrc_x; ix += 8) {
|
for (int ix = 0; ix < nrc_x; ix += k_nr) {
|
||||||
for (int k = 0; k < 8; ++k) x8[k] = (const block_q2_K *)((const char *)vx + (ix + k)*bx);
|
for (int k = 0; k < k_nr; ++k) x8[k] = (const block_q2_K *)((const char *)vx + (ix + k)*bx);
|
||||||
for (int i = 0; i < nb; ++i) {
|
for (int i = 0; i < nb; ++i) {
|
||||||
for (int k = 0; k < 8; ++k) {
|
for (int k = 0; k < k_nr; ++k) {
|
||||||
auto vd = _mm256_set1_ps(GGML_FP16_TO_FP32(x8[k][i].d));
|
auto vd = _mm256_set1_ps(GGML_FP16_TO_FP32(x8[k][i].d));
|
||||||
auto vm = _mm256_mul_ps(_mm256_set1_ps(GGML_FP16_TO_FP32(x8[k][i].dmin)), _mm256_set1_ps(-1.f));
|
auto vm = _mm256_mul_ps(_mm256_set1_ps(GGML_FP16_TO_FP32(x8[k][i].dmin)), _mm256_set1_ps(-1.f));
|
||||||
auto block_max = _mm256_setzero_ps();
|
auto block_max = _mm256_setzero_ps();
|
||||||
@@ -2136,13 +2143,18 @@ void iqk_convert_q2_k_q8_k_r8(int n, const void * vx, size_t bx, void * vy, int
|
|||||||
i0 = _mm256_permutevar8x32_epi32(i0, perm);
|
i0 = _mm256_permutevar8x32_epi32(i0, perm);
|
||||||
|
|
||||||
_mm256_storeu_si256((__m256i *)block, i0);
|
_mm256_storeu_si256((__m256i *)block, i0);
|
||||||
auto q8 = (uint32_t *)y[i].qs + 64*ib32;
|
auto q8 = (uint32_t *)y[i].qs + 8*k_nr*ib32;
|
||||||
for (int l = 0; l < 4; ++l) {
|
for (int l = 0; l < 8; ++l) {
|
||||||
q8[8*l + k + 0] = block[l + 0];
|
q8[k_nr*l + k] = block[l];
|
||||||
q8[8*l + k + 32] = block[l + 4];
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
#ifdef HAVE_FANCY_SIMD
|
||||||
|
for (int l = 0; l < 64; ++l) {
|
||||||
|
auto v = _mm512_xor_si512(_mm512_loadu_si512((const __m512i *)y[i].qs + l), _mm512_set1_epi8(-128));
|
||||||
|
_mm512_storeu_si512((__m512i *)y[i].qs + l, v);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
y += nb;
|
y += nb;
|
||||||
}
|
}
|
||||||
@@ -2458,14 +2470,21 @@ void iqk_convert_q3_k_q8_0_r8(int n, const void * vx, size_t bx, void * vy, int
|
|||||||
}
|
}
|
||||||
|
|
||||||
void iqk_convert_q3_k_q8_k_r8(int n, const void * vx, size_t bx, void * vy, int nrc_x) {
|
void iqk_convert_q3_k_q8_k_r8(int n, const void * vx, size_t bx, void * vy, int nrc_x) {
|
||||||
|
#ifdef HAVE_FANCY_SIMD
|
||||||
|
constexpr int k_nr = 16;
|
||||||
|
using block_q8_k_r = block_q8_k_r16;
|
||||||
|
#else
|
||||||
|
constexpr int k_nr = 8;
|
||||||
|
using block_q8_k_r = block_q8_k_r8;
|
||||||
|
#endif
|
||||||
GGML_ASSERT(n%QK_K == 0);
|
GGML_ASSERT(n%QK_K == 0);
|
||||||
GGML_ASSERT(nrc_x%8 == 0);
|
GGML_ASSERT(nrc_x%k_nr == 0);
|
||||||
|
|
||||||
int nb = n/QK_K;
|
int nb = n/QK_K;
|
||||||
|
|
||||||
const block_q3_K * x8[8];
|
const block_q3_K * x8[k_nr];
|
||||||
|
|
||||||
block_q8_k_r8 * y = (block_q8_k_r8 *)vy;
|
block_q8_k_r * y = (block_q8_k_r *)vy;
|
||||||
|
|
||||||
uint32_t block[8];
|
uint32_t block[8];
|
||||||
__m256i values[8];
|
__m256i values[8];
|
||||||
@@ -2476,10 +2495,10 @@ void iqk_convert_q3_k_q8_k_r8(int n, const void * vx, size_t bx, void * vy, int
|
|||||||
|
|
||||||
union { __m256i vec; int16_t val[16]; } helper;
|
union { __m256i vec; int16_t val[16]; } helper;
|
||||||
|
|
||||||
for (int ix = 0; ix < nrc_x; ix += 8) {
|
for (int ix = 0; ix < nrc_x; ix += k_nr) {
|
||||||
for (int k = 0; k < 8; ++k) x8[k] = (const block_q3_K *)((const char *)vx + (ix + k)*bx);
|
for (int k = 0; k < k_nr; ++k) x8[k] = (const block_q3_K *)((const char *)vx + (ix + k)*bx);
|
||||||
for (int i = 0; i < nb; ++i) {
|
for (int i = 0; i < nb; ++i) {
|
||||||
for (int k = 0; k < 8; ++k) {
|
for (int k = 0; k < k_nr; ++k) {
|
||||||
float d = GGML_FP16_TO_FP32(x8[k][i].d);
|
float d = GGML_FP16_TO_FP32(x8[k][i].d);
|
||||||
auto hbits = _mm256_loadu_si256((const __m256i *)x8[k][i].hmask);
|
auto hbits = _mm256_loadu_si256((const __m256i *)x8[k][i].hmask);
|
||||||
helper.vec = _mm256_cvtepi8_epi16(sc3.make_scales((const uint16_t *)x8[k][i].scales));
|
helper.vec = _mm256_cvtepi8_epi16(sc3.make_scales((const uint16_t *)x8[k][i].scales));
|
||||||
@@ -2549,12 +2568,18 @@ void iqk_convert_q3_k_q8_k_r8(int n, const void * vx, size_t bx, void * vy, int
|
|||||||
_mm_storeu_si128((__m128i *)block+0, _mm_unpacklo_epi64(i0_l, i0_h));
|
_mm_storeu_si128((__m128i *)block+0, _mm_unpacklo_epi64(i0_l, i0_h));
|
||||||
_mm_storeu_si128((__m128i *)block+1, _mm_unpackhi_epi64(i0_l, i0_h));
|
_mm_storeu_si128((__m128i *)block+1, _mm_unpackhi_epi64(i0_l, i0_h));
|
||||||
}
|
}
|
||||||
auto qs = (uint32_t *)y[i].qs + 64*ib32;
|
auto qs = (uint32_t *)y[i].qs + 8*k_nr*ib32;
|
||||||
for (int l = 0; l < 8; ++l) {
|
for (int l = 0; l < 8; ++l) {
|
||||||
qs[8*l + k] = block[l];
|
qs[k_nr*l + k] = block[l];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
#ifdef HAVE_FANCY_SIMD
|
||||||
|
for (int l = 0; l < 64; ++l) {
|
||||||
|
auto v = _mm512_xor_si512(_mm512_loadu_si512((const __m512i *)y[i].qs + l), _mm512_set1_epi8(-128));
|
||||||
|
_mm512_storeu_si512((__m512i *)y[i].qs + l, v);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
y += nb;
|
y += nb;
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user