mirror of
https://github.com/ikawrakow/ik_llama.cpp.git
synced 2026-04-27 01:49:28 +00:00
Q8_0_R4 (#120)
* Adding q8_0_r4 We get PP-512(LLaMA-3.1-8B) = 268 t/s on a Ryzen-7950X compared to 175.6 t/s for Q8_0. * q8_0_r4: NEON We get PP-512(LLaMA-3.1-8B) = 112.6 t/s on M2-Max. * q8_0_r4: Zen4 matrix-vector specialization --------- Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com>
This commit is contained in:
@@ -3287,3 +3287,81 @@ void vec_dot_q4_0_r4_q8_0(int n, float * s, size_t bs, const void * vx, size_t b
|
||||
GGML_UNUSED(by);
|
||||
}
|
||||
|
||||
|
||||
//
|
||||
// ========================================= q8_0_r4
|
||||
//
|
||||
void quantize_row_q8_0_r4_ref(const float * x, block_q8_0_x4 * y, int64_t k) {
|
||||
// we assume we are called with 4 rows
|
||||
quantize_q8_0_r4(x, (void *)y, 4, k/4, nullptr);
|
||||
}
|
||||
|
||||
void quantize_row_q8_0_r4(const float * x, void * y, int64_t k) {
|
||||
// we assume we are called with 4 rows
|
||||
quantize_q8_0_r4(x, y, 4, k/4, nullptr);
|
||||
}
|
||||
|
||||
static void repack_q8_0(int nrows, int n_per_row, const block_q8_0 * x, block_q8_0_x4 * y) {
|
||||
GGML_ASSERT(nrows%4 == 0);
|
||||
GGML_ASSERT(n_per_row%QK8_0 == 0);
|
||||
int nblock = n_per_row/QK8_0;
|
||||
const block_q8_0 * x4[4];
|
||||
for (int row = 0; row < nrows; row += 4) {
|
||||
for (int k = 0; k < 4; ++k) x4[k] = x + nblock*k;
|
||||
for (int ib = 0; ib < nblock; ++ib) {
|
||||
for (int k = 0; k < 4; ++k) y[ib].d[k] = x4[k][ib].d;
|
||||
for (int l = 0; l < 4; ++l) {
|
||||
for (int k = 0; k < 4; ++k) for (int i = 0; i < 4; ++i) {
|
||||
y[ib].qs[32*l+4*k+i+ 0] = x4[k][ib].qs[i+4*l+ 0];
|
||||
y[ib].qs[32*l+4*k+i+16] = x4[k][ib].qs[i+4*l+16];
|
||||
}
|
||||
}
|
||||
}
|
||||
x += 4*nblock;
|
||||
y += nblock;
|
||||
}
|
||||
}
|
||||
|
||||
size_t quantize_q8_0_r4(const float * src, void * dst, int64_t nrows, int64_t n_per_row, const float * imatrix) {
|
||||
GGML_ASSERT(nrows%4 == 0);
|
||||
auto row_size_0 = ggml_row_size(GGML_TYPE_Q8_0, n_per_row);
|
||||
std::vector<char> qtmp(4*row_size_0);
|
||||
char * qrow = (char *)dst;
|
||||
for (int row = 0; row < nrows; row += 4) {
|
||||
quantize_q8_0(src, qtmp.data(), 4, n_per_row, imatrix);
|
||||
repack_q8_0(4, n_per_row, (const block_q8_0 *)qtmp.data(), (block_q8_0_x4 *)qrow);
|
||||
src += 4*n_per_row;
|
||||
qrow += 4*row_size_0;
|
||||
}
|
||||
return nrows*row_size_0;
|
||||
}
|
||||
|
||||
void dequantize_row_q8_0_r4(const block_q8_0_x4 * x, float * y, int64_t k) {
|
||||
// we assume we are called with 4 rows
|
||||
int n_per_row = k/4;
|
||||
int nb = n_per_row/QK8_0;
|
||||
float * yk[4];
|
||||
for (int k = 0; k < 4; ++k) yk[k] = y + k*n_per_row;
|
||||
for (int ib = 0; ib < nb; ++ib) {
|
||||
for (int k = 0; k < 4; ++k) {
|
||||
float scale = GGML_FP16_TO_FP32(x[ib].d[k]);
|
||||
for (int l = 0; l < 4; ++l) for (int i = 0; i < 4; ++i) {
|
||||
yk[k][QK8_0*ib+4*l+i+ 0] = scale * x[ib].qs[QK8_0*l+4*k+i+ 0];
|
||||
yk[k][QK8_0*ib+4*l+i+16] = scale * x[ib].qs[QK8_0*l+4*k+i+16];
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void vec_dot_q8_0_r4_q8_0(int n, float * s, size_t bs, const void * vx, size_t bx, const void * vy, size_t by, int nrc) {
|
||||
#if GGML_USE_IQK_MULMAT
|
||||
if (iqk_mul_mat(1, 1, n, GGML_TYPE_Q8_0_R4, vx, 0, GGML_TYPE_Q8_0, vy, 0, s, 0, 0, 1)) {
|
||||
return;
|
||||
}
|
||||
#endif
|
||||
GGML_ASSERT(n%QK4_NL == 0);
|
||||
GGML_ASSERT(nrc == 1);
|
||||
GGML_UNUSED(bs);
|
||||
GGML_UNUSED(bx);
|
||||
GGML_UNUSED(by);
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user