Kawrakow
|
49662cbed3
ggml : SOTA 2-bit quants (add IQ2_XS) (#4856)
|
2 years ago |
Timothy Cronin
|
f85a973aa1
ggml : remove ggml_cpy_inplace and ggml_cont_inplace (ggml/693)
|
2 years ago |
Halalaluyafail3
|
c910e3c28a
Fix execlp call (ggml/689)
|
2 years ago |
Kawrakow
|
dd5ae06405
SOTA 2-bit quants (#4773)
|
2 years ago |
Georgi Gerganov
|
c1d7cb28d3
ggml : do not sched_yield when calling BLAS (#4761)
|
2 years ago |
Guillaume Wenzek
|
5f66ebca9c
ggml : extend ggml_get_rows, ggml_repeat, ggml_concat (ggml/639)
|
2 years ago |
automaticcat
|
24a447e20a
ggml : add ggml_cpu_has_avx_vnni() (#4589)
|
2 years ago |
bssrdf
|
afc8c19291
ggml : fix some mul mat cases + add tests for src1 F16 (ggml/669)
|
2 years ago |
slaren
|
dc68f0054c
cuda : fix vmm pool with multi GPU (#4620)
|
2 years ago |
WillCorticesAI
|
de8e496437
Update comment for AdamW implementation reference. (#4604)
|
2 years ago |
slaren
|
5bf3953d7e
cuda : improve cuda pool efficiency using virtual memory (#4606)
|
2 years ago |
slaren
|
48b7ff193e
llama : fix platforms without mmap (#4578)
|
2 years ago |
Herman Semenov
|
48b24b170e
ggml : add comment about backward GGML_OP_DIAG_MASK_INF (#4203)
|
2 years ago |
Georgi Gerganov
|
afefa319f1
ggml : change ggml_scale to take a float instead of tensor (#4573)
|
2 years ago |
slaren
|
d232aca5a7
llama : initial ggml-backend integration (#4520)
|
2 years ago |
Ebey Abraham
|
b9e74f9bca
llama : add phi-2 + fix NeoX rope + ggml_mul_mat_set_prec (#4490)
|
2 years ago |
slaren
|
ee4725a686
ggml : group mul_mat_id rows by matrix (cpu only) (#4480)
|
2 years ago |
slaren
|
6744dbe924
ggml : use ggml_row_size where possible (#4472)
|
2 years ago |
slaren
|
cafcd4f895
ggml : remove n_dims from ggml_tensor (#4469)
|
2 years ago |
LostRuins
|
20a68a7030
ggml : add ggml_row_size() (fixes llama out of space) (#4461)
|
2 years ago |
Georgi Gerganov
|
55e87c3749
ggml : fix OpenCL broadcast requirement for ggml_mul (close #4453)
|
2 years ago |
Georgi Gerganov
|
4d98d9a656
sync : ggml (SD ops, tests, kernels) (#4444)
|
2 years ago |
slaren
|
799a1cb13b
llama : add Mixtral support (#4406)
|
2 years ago |
Richard Kiss
|
9494d7c477
english : use `typos` to fix comments and logs (#4354)
|
2 years ago |
Georgi Gerganov
|
fe680e3d10
sync : ggml (new ops, tests, backend, etc.) (#4359)
|
2 years ago |
Georgi Gerganov
|
fbbc42827b
ggml : reuse ggml_get_n_tasks() in ggml_graph_plan() (#4308)
|
2 years ago |
Georgi Gerganov
|
adf3de4f69
ggml : fix soft max out-of-bounds access (#4307)
|
2 years ago |
Georgi Gerganov
|
ef47ec18da
ggml : add ggml_soft_max_ext (#4256)
|
2 years ago |
Georgi Gerganov
|
8406b0924b
ggml : re-enable BLAS for CPU when src0 != F32 + remove redundant full offload checks in llama.cpp (#4240)
|
2 years ago |
Jared Van Bortel
|
f3b269813f
ggml : fix -Warray-bounds warning with gcc (#4231)
|
2 years ago |