mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-25 02:44:36 +00:00
123 lines
5.7 KiB
C
123 lines
5.7 KiB
C
|
#pragma once
|
||
|
|
||
|
#include "ggml.h"
|
||
|
|
||
|
#include <stdint.h>
|
||
|
#include <assert.h>
|
||
|
#include <stddef.h>
|
||
|
|
||
|
// Super-block size
|
||
|
#define QK_K 256
|
||
|
|
||
|
//
|
||
|
// Super-block quantization structures
|
||
|
//
|
||
|
|
||
|
// 2-bit quantization
|
||
|
// weight is represented as x = a * q + b
|
||
|
// 16 blocks of 16 elemenets each
|
||
|
// Effectively 2.5625 bits per weight
|
||
|
typedef struct {
|
||
|
uint8_t scales[QK_K/16]; // scales and mins, quantized with 4 bits
|
||
|
uint8_t qs[QK_K/4]; // quants
|
||
|
ggml_fp16_t d; // super-block scale for quantized scales
|
||
|
ggml_fp16_t dmin; // super-block scale for quantized mins
|
||
|
} block_q2_k;
|
||
|
static_assert(sizeof(block_q2_k) == 2*sizeof(ggml_fp16_t) + QK_K/16 + QK_K/4, "wrong q2_k block size/padding");
|
||
|
|
||
|
// 3-bit quantization
|
||
|
// weight is represented as x = a * q
|
||
|
// 16 blocks of 16 elemenets each
|
||
|
// Effectively 3.4375 bits per weight
|
||
|
typedef struct {
|
||
|
uint8_t hmask[QK_K/8]; // quants - high bit
|
||
|
uint8_t qs[QK_K/4]; // quants - low 2 bits
|
||
|
uint8_t scales[3*QK_K/64]; // scales, quantized with 6 bits
|
||
|
ggml_fp16_t d; // super-block scale
|
||
|
} block_q3_k;
|
||
|
static_assert(sizeof(block_q3_k) == sizeof(ggml_fp16_t) + QK_K / 4 + 11 * QK_K / 64, "wrong q3_k block size/padding");
|
||
|
|
||
|
// 4-bit quantization
|
||
|
// 16 blocks of 32 elements each
|
||
|
// weight is represented as x = a * q + b
|
||
|
// Effectively 4.5 bits per weight
|
||
|
typedef struct {
|
||
|
ggml_fp16_t d; // super-block scale for quantized scales
|
||
|
ggml_fp16_t dmin; // super-block scale for quantized mins
|
||
|
uint8_t scales[3*QK_K/64]; // scales and mins, quantized with 6 bits
|
||
|
uint8_t qs[QK_K/2]; // 4--bit quants
|
||
|
} block_q4_k;
|
||
|
static_assert(sizeof(block_q4_k) == 2*sizeof(ggml_fp16_t) + 3*QK_K/64 + QK_K/2, "wrong q4_k block size/padding");
|
||
|
|
||
|
// 5-bit quantization
|
||
|
// 16 blocks of 32 elements each
|
||
|
// weight is represented as x = a * q + b
|
||
|
// Effectively 5.5 bits per weight
|
||
|
typedef struct {
|
||
|
ggml_fp16_t d; // super-block scale for quantized scales
|
||
|
ggml_fp16_t dmin; // super-block scale for quantized mins
|
||
|
uint8_t scales[3*QK_K/64]; // scales and mins, quantized with 6 bits
|
||
|
uint8_t qh[QK_K/8]; // quants, high bit
|
||
|
uint8_t qs[QK_K/2]; // quants, low 4 bits
|
||
|
} block_q5_k;
|
||
|
static_assert(sizeof(block_q5_k) == 2*sizeof(ggml_fp16_t) + 3*QK_K/64 + QK_K/2 + QK_K/8, "wrong q5_k block size/padding");
|
||
|
|
||
|
// 6-bit quantization
|
||
|
// weight is represented as x = a * q
|
||
|
// 16 blocks of 16 elemenets each
|
||
|
// Effectively 6.5625 bits per weight
|
||
|
typedef struct {
|
||
|
uint8_t ql[QK_K/2]; // quants, lower 4 bits
|
||
|
uint8_t qh[QK_K/4]; // quants, upper 2 bits
|
||
|
int8_t scales[QK_K/16]; // scales, quantized with 8 bits
|
||
|
ggml_fp16_t d; // super-block scale
|
||
|
} block_q6_k;
|
||
|
static_assert(sizeof(block_q6_k) == sizeof(ggml_fp16_t) + QK_K / 16 + 3*QK_K/4, "wrong q6_k block size/padding");
|
||
|
|
||
|
// This is only used for intermediate quantization and dot products
|
||
|
typedef struct {
|
||
|
float d; // delta
|
||
|
int8_t qs[QK_K]; // quants
|
||
|
int16_t bsums[QK_K/16]; // sum of quants in groups of 16
|
||
|
} block_q8_k;
|
||
|
static_assert(sizeof(block_q8_k) == sizeof(float) + QK_K + QK_K/16*sizeof(int16_t), "wrong q8_k block size/padding");
|
||
|
|
||
|
|
||
|
// Quantization
|
||
|
void quantize_row_q2_k_reference(const float * restrict x, block_q2_k * restrict y, int k);
|
||
|
void quantize_row_q3_k_reference(const float * restrict x, block_q3_k * restrict y, int k);
|
||
|
void quantize_row_q4_k_reference(const float * restrict x, block_q4_k * restrict y, int k);
|
||
|
void quantize_row_q5_k_reference(const float * restrict x, block_q5_k * restrict y, int k);
|
||
|
void quantize_row_q6_k_reference(const float * restrict x, block_q6_k * restrict y, int k);
|
||
|
void quantize_row_q8_k_reference(const float * restrict x, block_q8_k * restrict y, int k);
|
||
|
|
||
|
void quantize_row_q2_k(const float * restrict x, void * restrict y, int k);
|
||
|
void quantize_row_q3_k(const float * restrict x, void * restrict y, int k);
|
||
|
void quantize_row_q4_k(const float * restrict x, void * restrict y, int k);
|
||
|
void quantize_row_q5_k(const float * restrict x, void * restrict y, int k);
|
||
|
void quantize_row_q6_k(const float * restrict x, void * restrict y, int k);
|
||
|
void quantize_row_q8_k(const float * restrict x, void * restrict y, int k);
|
||
|
|
||
|
// Dequantization
|
||
|
void dequantize_row_q2_k(const block_q2_k * restrict x, float * restrict y, int k);
|
||
|
void dequantize_row_q3_k(const block_q3_k * restrict x, float * restrict y, int k);
|
||
|
void dequantize_row_q4_k(const block_q4_k * restrict x, float * restrict y, int k);
|
||
|
void dequantize_row_q5_k(const block_q5_k * restrict x, float * restrict y, int k);
|
||
|
void dequantize_row_q6_k(const block_q6_k * restrict x, float * restrict y, int k);
|
||
|
void dequantize_row_q8_k(const block_q8_k * restrict x, float * restrict y, int k);
|
||
|
|
||
|
// Dot product
|
||
|
void ggml_vec_dot_q2_k_q8_k(int n, float * restrict s, const void * restrict vx, const void * restrict vy);
|
||
|
void ggml_vec_dot_q3_k_q8_k(int n, float * restrict s, const void * restrict vx, const void * restrict vy);
|
||
|
void ggml_vec_dot_q4_k_q8_k(int n, float * restrict s, const void * restrict vx, const void * restrict vy);
|
||
|
void ggml_vec_dot_q5_k_q8_k(int n, float * restrict s, const void * restrict vx, const void * restrict vy);
|
||
|
void ggml_vec_dot_q6_k_q8_k(int n, float * restrict s, const void * restrict vx, const void * restrict vy);
|
||
|
|
||
|
// Quantization with histogram collection
|
||
|
size_t ggml_quantize_q2_k(const float * src, void * dst, int n, int k, int64_t * hist);
|
||
|
size_t ggml_quantize_q3_k(const float * src, void * dst, int n, int k, int64_t * hist);
|
||
|
size_t ggml_quantize_q4_k(const float * src, void * dst, int n, int k, int64_t * hist);
|
||
|
size_t ggml_quantize_q5_k(const float * src, void * dst, int n, int k, int64_t * hist);
|
||
|
size_t ggml_quantize_q6_k(const float * src, void * dst, int n, int k, int64_t * hist);
|
||
|
|