whisper.cpp / ggml

Commit History

ggml : add logging for native build options/vars (#2935)
aaf8a91
unverified

danbev commited on

examples : command.wasm updates (#2904)
0db3249
unverified

danbev commited on

cmake : fix ggml-config (ggml/0)
40f0325

ggerganov commited on

ggml-cpu: faster AVX2 variant for IQ1_M (llama/12216)
591cbfb

Rémy O commited on

metal : simplify kernel arguments using a struct (ggml/3229) (llama/12194)
092277a

BB-fat alexju commited on

metal : fix default.metallib build (llama/12224)
838efb6

danbev commited on

opencl: Noncontiguous `norm`, `rms_norm`, disable `fp16` for some ops (llama/12217)
94449e3

lhez commited on

cmake : fix undefined reference errors for std::filesystem in ggml (#12092) (llama/12094)
dc68418

xiaofei Ray Lee commited on

CUDA: fix FA logic for PTX 7.0 and CC >= 7.5 (llama/12222)
4dc8a81

JohannesGaessler commited on

HIP/CUDA: set the paramerter value in maintain_cuda_graph instead of replaceing it. (llama/12209)
18afa4b

uvos commited on

opencl : fix buffer alignment (llama/12197)
7d25156

linehill commited on

opencl : fix `ulong` kernel args were set from `int` variables (llama/12174)
67ffff0

linehill commited on

opencl : fix profile-related errors (llama/12095)
e11a847

simon886212 ubuntu commited on

ggml-cpu: Faster IQ1 mul_mat_vec on AVX2 using BMI2 instructions (llama/12154)
05466a9

Rémy O commited on

SYCL: Disable f16 Unary OPs as not supported by the kernels (llama/12201)
723b8b4

Akarshan Biswas commited on

ggml : fix GGMLMetalClass ODR (llama/12200)
2094cb7

pacominev commited on

ggml : ggml_compute_forward_concat() for arbitrary tensor type (ggml/1118)
c9a49f9

vmobilis commited on

vulkan : sync (llama/0)
4c17fa1

ggerganov commited on

ggml : portability fixes for VS 2017 (llama/12150)
49e3343

mgroeber9110 Marcus Groeber commited on

HIP: implement FlashAttention via rocWMMA for CDNA and RDNA3+ (llama/12032)
a027c1d

David Huang commited on

ggml : fix kleidiai build (llama/12159)
dbc0180

ag2s20150909 commited on

SYCL: Move CPY kernels to a separate file and add few missing kernels (llama/12133)
1d6d451

Akarshan Biswas commited on

ggml-backend : keep paths in native string type when possible (llama/12144)
6e89d8c

Diego Devesa commited on

CUDA: compress mode option and default to size (llama/12029)
4ec988a

Erik Scholz commited on

ggml : upgrade init_tensor API to return a ggml_status (llama/11854)
d6b6852

William Tambellini slaren commited on

vulkan: add specific MMV kernels for IQ2 and IQ3 quants + optimizations (llama/11595)
d7d82b9

Rémy O commited on

CUDA: fix logic for V100 + GGML_CUDA_FORCE_MMQ (llama/12098)
0b52fcc

JohannesGaessler commited on

ggml: aarch64: implement SVE kernels for q2_k_q8_k vector dot (llama/12064)
459beb1

Prashant Vithule vithulep commited on

CANN: Fix build error with GCC 13 (llama/11990)
dcf68db

hipudding commited on

vulkan: matmul dequantization improvements (llama/12015)
ffdf466

Eve commited on

vulkan: improve im2col (llama/11826)
f6cff0a

Daniele commited on

cmake: Fix ggml backend dependencies and installation (llama/11818)
c6c2a2c

Vladimir Vuksanovic commited on

vulkan: fix assertion when qy_needs_dequant (llama/12068)
271c7e4

jeffbolznv commited on

ggml-cpu: Fix build with sve (llama/12059)
4be146e

mollysama commited on

cuda: unary ops as float + de-duplicate (ggml/1130)
4bec2e4

cmdr2 commited on

cuda/vulkan: specify fp32-only support for some operations in supports_op (ggml/1129)
f959b90

cmdr2 commited on

cuda/cpu: Increase support for fp16 unary operations (ggml/1125)
67e8c32

cmdr2 commited on

Told cmake to install ggml-cpp.h as a public header file. (ggml/1126)
3d4f29c

petterreinholdtsen Petter Reinholdtsen commited on

whisper : support GGML_BACKEND_DL (#2843)
2e6437e
unverified

Diego Devesa ggerganov commited on

Support pure float16 add/sub/mul/div operations in the CUDA (and CPU) backend (ggml/1121)
2b94a24

cmdr2 commited on

metal : copy kernels for quant to F32/F16 conversions (llama/12017)
6c8e7ec

Garf ggerganov commited on

opencl: fix for small models (llama/11950)
4532dc6

lhez Shawn Gu Skyler Szot commited on

Optimize mul_mat for Q4_0 on Intel GPU (llama/12035)
14fd317

Neo Zhang Jianyu arthw commited on

SYCL: Fix GGML_SYCL_DEBUG macro (llama/11995)
310a36c

qnixsynapse commited on

ggml-cpu: Support s390x SIMD Instruction Set (llama/12019)
4aa54ec

Aaron Teo Jinyang He junchao-zhao commited on

CUDA: app option to compile without FlashAttention (llama/12025)
fbc5f16

JohannesGaessler commited on

CUDA: optimize FA for GQA + large batches (llama/12014)
6662d54

JohannesGaessler commited on

cuda: Add Q5_1, Q5_0, Q4_1 and Q4_0 to F32 conversion support. (llama/12000)
6cb8158

Garf commited on

CUDA: correct the lowest Maxwell supported by CUDA 12 (llama/11984)
6641178

PureJourney JohannesGaessler commited on

MUSA: support ARM64 and enable dp4a .etc (llama/11843)
ab96dac

Bodhi Bodhi Hu commited on