Releases: ggml-org/llama.cpp
Releases · ggml-org/llama.cpp
b5606
b5604
SYCL: Implement few same quantized type copy kernels (#13739) * SYCL: Implement few same quantized type copy kernels * Use memcpy for copying contiguous tensors ggml-ci * feat(sycl): add contiguous tensor copy support and device checks Adds a memcpy path for contiguous tensors of the same type to optimize data transfer. Updates device support checks to recognize contiguous tensor operations, improving compatibility and performance. * refactor: replace specific block copy functions with template The changes replace multiple redundant block copy functions (e.g., cpy_block_q8_0_q8_0, cpy_block_q5_0_q5_0) with a single templated function cpy_blck_q_q. This reduces code duplication by using a generic template that works for any block type, improving maintainability while preserving the same functionality. The template is instantiated with specific block types (e.g., block_q8_0) where needed. * Exclude BF16 support for COPY tensors for now ggml-ci * perf: adjust SYCL copy kernel block sizes for efficiency Use ceil_div to ensure full element coverage and update nd_range parameters to better align with SYCL block sizes, improving parallelism and device utilization in copy operations.
b5603
llama : fix llama_model_chat_template with template name (LLM_KV with…
b5602
llama : deprecate llama_kv_self_ API (#14030) * llama : deprecate llama_kv_self_ API ggml-ci * llama : allow llama_memory_(nullptr) ggml-ci * memory : add flag for optional data clear in llama_memory_clear ggml-ci
b5601
context : fix SWA-related warning for multiple sequences (#14045)
b5600
llama : support multiple classifier outputs and labels (#13940)
b5598
vulkan: Enable VK_KHR_cooperative_matrix extension for Intel Xe2 GPUs…
b5596
memory : migrate from llama_kv_cache to more generic llama_memory (#1…
b5595
llama : allow using mmap without PrefetchVirtualMemory, apply GGML_WI…
b5593
vocab : warn about missing mask token (#14022)