|
|
@@ -1,48 +1,37 @@
|
|
|
#include "common.cuh"
|
|
|
|
|
|
-static __device__ __forceinline__ void dequantize_q4_0(const void * vx, const int64_t ib, const int iqs, dfloat2 & v){
|
|
|
+static __device__ __forceinline__ void dequantize_q4_0(const void * vx, const int64_t ib, const int iqs, float2 & v){
|
|
|
const block_q4_0 * x = (const block_q4_0 *) vx;
|
|
|
|
|
|
- const dfloat d = x[ib].d;
|
|
|
+ const float d = x[ib].d;
|
|
|
|
|
|
const int vui = x[ib].qs[iqs];
|
|
|
|
|
|
v.x = vui & 0xF;
|
|
|
v.y = vui >> 4;
|
|
|
|
|
|
-#ifdef GGML_CUDA_F16
|
|
|
- v = __hsub2(v, {8.0f, 8.0f});
|
|
|
- v = __hmul2(v, {d, d});
|
|
|
-#else
|
|
|
v.x = (v.x - 8.0f) * d;
|
|
|
v.y = (v.y - 8.0f) * d;
|
|
|
-#endif // GGML_CUDA_F16
|
|
|
}
|
|
|
|
|
|
-static __device__ __forceinline__ void dequantize_q4_1(const void * vx, const int64_t ib, const int iqs, dfloat2 & v){
|
|
|
+static __device__ __forceinline__ void dequantize_q4_1(const void * vx, const int64_t ib, const int iqs, float2 & v){
|
|
|
const block_q4_1 * x = (const block_q4_1 *) vx;
|
|
|
|
|
|
- const dfloat d = __low2half(x[ib].dm);
|
|
|
- const dfloat m = __high2half(x[ib].dm);
|
|
|
+ const float2 dm = __half22float2(x[ib].dm);
|
|
|
|
|
|
const int vui = x[ib].qs[iqs];
|
|
|
|
|
|
v.x = vui & 0xF;
|
|
|
v.y = vui >> 4;
|
|
|
|
|
|
-#ifdef GGML_CUDA_F16
|
|
|
- v = __hmul2(v, {d, d});
|
|
|
- v = __hadd2(v, {m, m});
|
|
|
-#else
|
|
|
- v.x = (v.x * d) + m;
|
|
|
- v.y = (v.y * d) + m;
|
|
|
-#endif // GGML_CUDA_F16
|
|
|
+ v.x = (v.x * dm.x) + dm.y;
|
|
|
+ v.y = (v.y * dm.x) + dm.y;
|
|
|
}
|
|
|
|
|
|
-static __device__ __forceinline__ void dequantize_q5_0(const void * vx, const int64_t ib, const int iqs, dfloat2 & v){
|
|
|
+static __device__ __forceinline__ void dequantize_q5_0(const void * vx, const int64_t ib, const int iqs, float2 & v){
|
|
|
const block_q5_0 * x = (const block_q5_0 *) vx;
|
|
|
|
|
|
- const dfloat d = x[ib].d;
|
|
|
+ const float d = x[ib].d;
|
|
|
|
|
|
uint32_t qh;
|
|
|
memcpy(&qh, x[ib].qh, sizeof(qh));
|
|
|
@@ -53,20 +42,14 @@ static __device__ __forceinline__ void dequantize_q5_0(const void * vx, const in
|
|
|
v.x = ((x[ib].qs[iqs] & 0xf) | xh_0);
|
|
|
v.y = ((x[ib].qs[iqs] >> 4) | xh_1);
|
|
|
|
|
|
-#ifdef GGML_CUDA_F16
|
|
|
- v = __hsub2(v, {16.0f, 16.0f});
|
|
|
- v = __hmul2(v, {d, d});
|
|
|
-#else
|
|
|
v.x = (v.x - 16.0f) * d;
|
|
|
v.y = (v.y - 16.0f) * d;
|
|
|
-#endif // GGML_CUDA_F16
|
|
|
}
|
|
|
|
|
|
-static __device__ __forceinline__ void dequantize_q5_1(const void * vx, const int64_t ib, const int iqs, dfloat2 & v){
|
|
|
+static __device__ __forceinline__ void dequantize_q5_1(const void * vx, const int64_t ib, const int iqs, float2 & v){
|
|
|
const block_q5_1 * x = (const block_q5_1 *) vx;
|
|
|
|
|
|
- const dfloat d = __low2half(x[ib].dm);
|
|
|
- const dfloat m = __high2half(x[ib].dm);
|
|
|
+ const float2 dm = __half22float2(x[ib].dm);
|
|
|
|
|
|
uint32_t qh;
|
|
|
memcpy(&qh, x[ib].qh, sizeof(qh));
|
|
|
@@ -77,27 +60,18 @@ static __device__ __forceinline__ void dequantize_q5_1(const void * vx, const in
|
|
|
v.x = ((x[ib].qs[iqs] & 0xf) | xh_0);
|
|
|
v.y = ((x[ib].qs[iqs] >> 4) | xh_1);
|
|
|
|
|
|
-#ifdef GGML_CUDA_F16
|
|
|
- v = __hmul2(v, {d, d});
|
|
|
- v = __hadd2(v, {m, m});
|
|
|
-#else
|
|
|
- v.x = (v.x * d) + m;
|
|
|
- v.y = (v.y * d) + m;
|
|
|
-#endif // GGML_CUDA_F16
|
|
|
+ v.x = (v.x * dm.x) + dm.y;
|
|
|
+ v.y = (v.y * dm.x) + dm.y;
|
|
|
}
|
|
|
|
|
|
-static __device__ __forceinline__ void dequantize_q8_0(const void * vx, const int64_t ib, const int iqs, dfloat2 & v){
|
|
|
+static __device__ __forceinline__ void dequantize_q8_0(const void * vx, const int64_t ib, const int iqs, float2 & v){
|
|
|
const block_q8_0 * x = (const block_q8_0 *) vx;
|
|
|
|
|
|
- const dfloat d = x[ib].d;
|
|
|
+ const float d = x[ib].d;
|
|
|
|
|
|
v.x = x[ib].qs[iqs + 0];
|
|
|
v.y = x[ib].qs[iqs + 1];
|
|
|
|
|
|
-#ifdef GGML_CUDA_F16
|
|
|
- v = __hmul2(v, {d, d});
|
|
|
-#else
|
|
|
v.x *= d;
|
|
|
v.y *= d;
|
|
|
-#endif // GGML_CUDA_F16
|
|
|
}
|